3.8 KiB
3.8 KiB
Google Gemini Model IDs
List of available model IDs for the Google Gemini API.
Last Updated: 2025-11-24
Model List
While there are many models available, gemini-2.5-flash is generally recommended for development at this time. It offers a good balance of cost and performance for a wide range of use cases.
Gemini 3.x (Latest)
| Model ID | Context | Max Output | Use Case |
|---|---|---|---|
google/gemini-3-pro-preview |
- | 64K | Latest high-performance model |
google/gemini-3-pro-image-preview |
- | - | Image generation |
google/gemini-3-pro-image-preview-edit |
- | - | Image editing |
Gemini 2.5
| Model ID | Context | Max Output | Use Case |
|---|---|---|---|
google/gemini-2.5-pro |
1M (2M planned) | - | High performance |
gemini-2.5-flash |
1M | - | Fast balanced model (recommended) |
gemini-2.5-flash-lite |
1M | - | Lightweight and fast |
Note: Free tier is limited to approximately 32K tokens. Gemini Advanced (2.5 Pro) supports 1M tokens.
Gemini 2.0
| Model ID | Context | Max Output | Use Case |
|---|---|---|---|
gemini-2.0-flash |
1M | - | Stable version |
Basic Usage
from langchain_google_genai import ChatGoogleGenerativeAI
# Recommended: Balanced model
llm = ChatGoogleGenerativeAI(model="gemini-2.5-flash")
# Also works with prefix
llm = ChatGoogleGenerativeAI(model="models/gemini-2.5-flash")
# High-performance version
llm = ChatGoogleGenerativeAI(model="google/gemini-3-pro")
# Lightweight version
llm = ChatGoogleGenerativeAI(model="gemini-2.5-flash-lite")
Environment Variables
export GOOGLE_API_KEY="your-api-key"
Model Selection Guide
| Use Case | Recommended Model |
|---|---|
| Cost-focused | gemini-2.5-flash-lite |
| Balanced | gemini-2.5-flash |
| Performance-focused | google/gemini-3-pro |
| Large context | gemini-2.5-pro (1M tokens) |
Gemini Features
1. Large Context Window
Gemini is the industry's first model to support 1M tokens:
| Tier | Context Limit |
|---|---|
| Gemini Advanced (2.5 Pro) | 1M tokens |
| Vertex AI | 1M tokens |
| Free tier | ~32K tokens |
Use Cases:
- Long document analysis
- Understanding entire codebases
- Long conversation history
# Processing large context
llm = ChatGoogleGenerativeAI(
model="gemini-2.5-pro",
max_tokens=8192 # Specify output token count
)
Future: Gemini 2.5 Pro is planned to support 2M token context windows.
2. Multimodal Support
Image input and generation capabilities (see Advanced Features for details).
Important Notes
- ❌ Deprecated: Gemini 1.0, 1.5 series are no longer available
- ✅ Migration Recommended: Use
gemini-2.5-flashor later models
Detailed Documentation
For advanced configuration and multimodal features, see: