Happy Horse 1.0 is now on ModelsLab

Try Now
Skip to main content
Available now on ModelsLab · Language Model

Google: Gemini 2.5 Flash Lite Preview 09-2025Flash Reasoning, Lite Latency

Run Multimodal Tasks Fast

1M Context

Process Text Video Audio

Handles text, image, video, audio, PDF inputs with 1,048,576 token limit.

Thinking Mode

Enable Step-by-Step Reasoning

Activates multi-pass thinking for complex logic and code execution.

Tool Integration

Function Calling Caching

Supports function calling, search grounding, structured outputs, batch API.

Examples

See what Google: Gemini 2.5 Flash Lite Preview 09-2025 can create

Copy any prompt below and try it yourself in the playground.

Code Debugger

Analyze this Python code for bugs: def factorial(n): if n == 0: return 1 else: return n * factorial(n-1). Fix errors and optimize for large n using memoization.

Document Summary

Summarize key points from this PDF contract uploaded as file. Highlight obligations, timelines, and payment terms in bullet points.

Image Analysis

Describe this architectural blueprint image. Identify materials, dimensions, and suggest structural improvements for earthquake resistance.

Video Transcript

Transcribe audio from this product demo video. Extract technical specs, features, and generate a comparison table with competitors.

For Developers

A few lines of code.
Reasoning LLM. One Call.

ModelsLab handles the infrastructure: fast inference, auto-scaling, and a developer-friendly API. No GPU management needed.

  • Serverless: scales to zero, scales to millions
  • Pay per token, no minimums
  • Python and JavaScript SDKs, plus REST API
import requests
response = requests.post(
"https://modelslab.com/api/v7/llm/chat/completions",
json={
"key": "YOUR_API_KEY",
"prompt": "",
"model_id": ""
}
)
print(response.json())

FAQ

Common questions about Google: Gemini 2.5 Flash Lite Preview 09-2025

Read the docs

Lightweight reasoning model from Google optimized for low latency. Supports 1M input tokens, multimodal inputs. Released September 25, 2025.

$0.10 per 1M input tokens, $0.40 per 1M output tokens. Cache read at $0.025 per 1M. Batch API supported.

Yes, processes text, image, video, audio, PDF. Outputs text only. Max output 65,536 tokens.

Includes thinking mode, function calling, code execution, structured outputs. Caching and file search enabled.

Stable version gemini-2.5-flash-lite available. Preview shuts down; use via APIs like ModelsLab for continuity.

1,048,576 input tokens standard. Knowledge cutoff January 2025. Thinking budget configurable.

Ready to create?

Start generating with Google: Gemini 2.5 Flash Lite Preview 09-2025 on ModelsLab.