116 lines
3.8 KiB
Markdown
116 lines
3.8 KiB
Markdown
# Google Gemini Model IDs
|
|
|
|
List of available model IDs for the Google Gemini API.
|
|
|
|
> **Last Updated**: 2025-11-24
|
|
|
|
## Model List
|
|
|
|
While there are many models available, `gemini-2.5-flash` is generally recommended for development at this time. It offers a good balance of cost and performance for a wide range of use cases.
|
|
|
|
### Gemini 3.x (Latest)
|
|
|
|
| Model ID | Context | Max Output | Use Case |
|
|
| ---------------------------------------- | ------------ | -------- | ------------------ |
|
|
| `google/gemini-3-pro-preview` | - | 64K | Latest high-performance model |
|
|
| `google/gemini-3-pro-image-preview` | - | - | Image generation |
|
|
| `google/gemini-3-pro-image-preview-edit` | - | - | Image editing |
|
|
|
|
### Gemini 2.5
|
|
|
|
| Model ID | Context | Max Output | Use Case |
|
|
| ----------------------- | ------------ | -------- | ---------------------- |
|
|
| `google/gemini-2.5-pro` | 1M (2M planned) | - | High performance |
|
|
| `gemini-2.5-flash` | 1M | - | Fast balanced model (recommended) |
|
|
| `gemini-2.5-flash-lite` | 1M | - | Lightweight and fast |
|
|
|
|
**Note**: Free tier is limited to approximately 32K tokens. Gemini Advanced (2.5 Pro) supports 1M tokens.
|
|
|
|
### Gemini 2.0
|
|
|
|
| Model ID | Context | Max Output | Use Case |
|
|
| ------------------ | ------------ | -------- | ------ |
|
|
| `gemini-2.0-flash` | 1M | - | Stable version |
|
|
|
|
## Basic Usage
|
|
|
|
```python
|
|
from langchain_google_genai import ChatGoogleGenerativeAI
|
|
|
|
# Recommended: Balanced model
|
|
llm = ChatGoogleGenerativeAI(model="gemini-2.5-flash")
|
|
|
|
# Also works with prefix
|
|
llm = ChatGoogleGenerativeAI(model="models/gemini-2.5-flash")
|
|
|
|
# High-performance version
|
|
llm = ChatGoogleGenerativeAI(model="google/gemini-3-pro")
|
|
|
|
# Lightweight version
|
|
llm = ChatGoogleGenerativeAI(model="gemini-2.5-flash-lite")
|
|
```
|
|
|
|
### Environment Variables
|
|
|
|
```bash
|
|
export GOOGLE_API_KEY="your-api-key"
|
|
```
|
|
|
|
## Model Selection Guide
|
|
|
|
| Use Case | Recommended Model |
|
|
| ------------------ | ------------------------------ |
|
|
| Cost-focused | `gemini-2.5-flash-lite` |
|
|
| Balanced | `gemini-2.5-flash` |
|
|
| Performance-focused | `google/gemini-3-pro` |
|
|
| Large context | `gemini-2.5-pro` (1M tokens) |
|
|
|
|
## Gemini Features
|
|
|
|
### 1. Large Context Window
|
|
|
|
Gemini is the **industry's first model to support 1M tokens**:
|
|
|
|
| Tier | Context Limit |
|
|
| ------------------------- | ---------------- |
|
|
| Gemini Advanced (2.5 Pro) | 1M tokens |
|
|
| Vertex AI | 1M tokens |
|
|
| Free tier | ~32K tokens |
|
|
|
|
**Use Cases**:
|
|
|
|
- Long document analysis
|
|
- Understanding entire codebases
|
|
- Long conversation history
|
|
|
|
```python
|
|
# Processing large context
|
|
llm = ChatGoogleGenerativeAI(
|
|
model="gemini-2.5-pro",
|
|
max_tokens=8192 # Specify output token count
|
|
)
|
|
```
|
|
|
|
**Future**: Gemini 2.5 Pro is planned to support 2M token context windows.
|
|
|
|
### 2. Multimodal Support
|
|
|
|
Image input and generation capabilities (see [Advanced Features](06_llm_model_ids_gemini_advanced.md) for details).
|
|
|
|
## Important Notes
|
|
|
|
- ❌ **Deprecated**: Gemini 1.0, 1.5 series are no longer available
|
|
- ✅ **Migration Recommended**: Use `gemini-2.5-flash` or later models
|
|
|
|
## Detailed Documentation
|
|
|
|
For advanced configuration and multimodal features, see:
|
|
|
|
- **[Gemini Advanced Features](06_llm_model_ids_gemini_advanced.md)**
|
|
|
|
## Reference Links
|
|
|
|
- [Gemini API Official](https://ai.google.dev/gemini-api/docs/models)
|
|
- [Google AI Studio](https://makersuite.google.com/)
|
|
- [LangChain Integration](https://docs.langchain.com/oss/python/integrations/chat/google_generative_ai)
|