Initial commit
This commit is contained in:
115
skills/langgraph-master/06_llm_model_ids_gemini.md
Normal file
115
skills/langgraph-master/06_llm_model_ids_gemini.md
Normal file
@@ -0,0 +1,115 @@
|
||||
# Google Gemini Model IDs
|
||||
|
||||
List of available model IDs for the Google Gemini API.
|
||||
|
||||
> **Last Updated**: 2025-11-24
|
||||
|
||||
## Model List
|
||||
|
||||
While there are many models available, `gemini-2.5-flash` is generally recommended for development at this time. It offers a good balance of cost and performance for a wide range of use cases.
|
||||
|
||||
### Gemini 3.x (Latest)
|
||||
|
||||
| Model ID | Context | Max Output | Use Case |
|
||||
| ---------------------------------------- | ------------ | -------- | ------------------ |
|
||||
| `google/gemini-3-pro-preview` | - | 64K | Latest high-performance model |
|
||||
| `google/gemini-3-pro-image-preview` | - | - | Image generation |
|
||||
| `google/gemini-3-pro-image-preview-edit` | - | - | Image editing |
|
||||
|
||||
### Gemini 2.5
|
||||
|
||||
| Model ID | Context | Max Output | Use Case |
|
||||
| ----------------------- | ------------ | -------- | ---------------------- |
|
||||
| `google/gemini-2.5-pro` | 1M (2M planned) | - | High performance |
|
||||
| `gemini-2.5-flash` | 1M | - | Fast balanced model (recommended) |
|
||||
| `gemini-2.5-flash-lite` | 1M | - | Lightweight and fast |
|
||||
|
||||
**Note**: Free tier is limited to approximately 32K tokens. Gemini Advanced (2.5 Pro) supports 1M tokens.
|
||||
|
||||
### Gemini 2.0
|
||||
|
||||
| Model ID | Context | Max Output | Use Case |
|
||||
| ------------------ | ------------ | -------- | ------ |
|
||||
| `gemini-2.0-flash` | 1M | - | Stable version |
|
||||
|
||||
## Basic Usage
|
||||
|
||||
```python
|
||||
from langchain_google_genai import ChatGoogleGenerativeAI
|
||||
|
||||
# Recommended: Balanced model
|
||||
llm = ChatGoogleGenerativeAI(model="gemini-2.5-flash")
|
||||
|
||||
# Also works with prefix
|
||||
llm = ChatGoogleGenerativeAI(model="models/gemini-2.5-flash")
|
||||
|
||||
# High-performance version
|
||||
llm = ChatGoogleGenerativeAI(model="google/gemini-3-pro")
|
||||
|
||||
# Lightweight version
|
||||
llm = ChatGoogleGenerativeAI(model="gemini-2.5-flash-lite")
|
||||
```
|
||||
|
||||
### Environment Variables
|
||||
|
||||
```bash
|
||||
export GOOGLE_API_KEY="your-api-key"
|
||||
```
|
||||
|
||||
## Model Selection Guide
|
||||
|
||||
| Use Case | Recommended Model |
|
||||
| ------------------ | ------------------------------ |
|
||||
| Cost-focused | `gemini-2.5-flash-lite` |
|
||||
| Balanced | `gemini-2.5-flash` |
|
||||
| Performance-focused | `google/gemini-3-pro` |
|
||||
| Large context | `gemini-2.5-pro` (1M tokens) |
|
||||
|
||||
## Gemini Features
|
||||
|
||||
### 1. Large Context Window
|
||||
|
||||
Gemini is the **industry's first model to support 1M tokens**:
|
||||
|
||||
| Tier | Context Limit |
|
||||
| ------------------------- | ---------------- |
|
||||
| Gemini Advanced (2.5 Pro) | 1M tokens |
|
||||
| Vertex AI | 1M tokens |
|
||||
| Free tier | ~32K tokens |
|
||||
|
||||
**Use Cases**:
|
||||
|
||||
- Long document analysis
|
||||
- Understanding entire codebases
|
||||
- Long conversation history
|
||||
|
||||
```python
|
||||
# Processing large context
|
||||
llm = ChatGoogleGenerativeAI(
|
||||
model="gemini-2.5-pro",
|
||||
max_tokens=8192 # Specify output token count
|
||||
)
|
||||
```
|
||||
|
||||
**Future**: Gemini 2.5 Pro is planned to support 2M token context windows.
|
||||
|
||||
### 2. Multimodal Support
|
||||
|
||||
Image input and generation capabilities (see [Advanced Features](06_llm_model_ids_gemini_advanced.md) for details).
|
||||
|
||||
## Important Notes
|
||||
|
||||
- ❌ **Deprecated**: Gemini 1.0, 1.5 series are no longer available
|
||||
- ✅ **Migration Recommended**: Use `gemini-2.5-flash` or later models
|
||||
|
||||
## Detailed Documentation
|
||||
|
||||
For advanced configuration and multimodal features, see:
|
||||
|
||||
- **[Gemini Advanced Features](06_llm_model_ids_gemini_advanced.md)**
|
||||
|
||||
## Reference Links
|
||||
|
||||
- [Gemini API Official](https://ai.google.dev/gemini-api/docs/models)
|
||||
- [Google AI Studio](https://makersuite.google.com/)
|
||||
- [LangChain Integration](https://docs.langchain.com/oss/python/integrations/chat/google_generative_ai)
|
||||
Reference in New Issue
Block a user