Files
gh-hiroshi75-protografico-p…/skills/langgraph-master/06_llm_model_ids_gemini.md
2025-11-29 18:45:58 +08:00

3.8 KiB

Google Gemini Model IDs

List of available model IDs for the Google Gemini API.

Last Updated: 2025-11-24

Model List

While there are many models available, gemini-2.5-flash is generally recommended for development at this time. It offers a good balance of cost and performance for a wide range of use cases.

Gemini 3.x (Latest)

Model ID Context Max Output Use Case
google/gemini-3-pro-preview - 64K Latest high-performance model
google/gemini-3-pro-image-preview - - Image generation
google/gemini-3-pro-image-preview-edit - - Image editing

Gemini 2.5

Model ID Context Max Output Use Case
google/gemini-2.5-pro 1M (2M planned) - High performance
gemini-2.5-flash 1M - Fast balanced model (recommended)
gemini-2.5-flash-lite 1M - Lightweight and fast

Note: Free tier is limited to approximately 32K tokens. Gemini Advanced (2.5 Pro) supports 1M tokens.

Gemini 2.0

Model ID Context Max Output Use Case
gemini-2.0-flash 1M - Stable version

Basic Usage

from langchain_google_genai import ChatGoogleGenerativeAI

# Recommended: Balanced model
llm = ChatGoogleGenerativeAI(model="gemini-2.5-flash")

# Also works with prefix
llm = ChatGoogleGenerativeAI(model="models/gemini-2.5-flash")

# High-performance version
llm = ChatGoogleGenerativeAI(model="google/gemini-3-pro")

# Lightweight version
llm = ChatGoogleGenerativeAI(model="gemini-2.5-flash-lite")

Environment Variables

export GOOGLE_API_KEY="your-api-key"

Model Selection Guide

Use Case Recommended Model
Cost-focused gemini-2.5-flash-lite
Balanced gemini-2.5-flash
Performance-focused google/gemini-3-pro
Large context gemini-2.5-pro (1M tokens)

Gemini Features

1. Large Context Window

Gemini is the industry's first model to support 1M tokens:

Tier Context Limit
Gemini Advanced (2.5 Pro) 1M tokens
Vertex AI 1M tokens
Free tier ~32K tokens

Use Cases:

  • Long document analysis
  • Understanding entire codebases
  • Long conversation history
# Processing large context
llm = ChatGoogleGenerativeAI(
    model="gemini-2.5-pro",
    max_tokens=8192  # Specify output token count
)

Future: Gemini 2.5 Pro is planned to support 2M token context windows.

2. Multimodal Support

Image input and generation capabilities (see Advanced Features for details).

Important Notes

  • Deprecated: Gemini 1.0, 1.5 series are no longer available
  • Migration Recommended: Use gemini-2.5-flash or later models

Detailed Documentation

For advanced configuration and multimodal features, see: