FlagshipGoogle

Gemini 3.1 Pro

Gemini 3.1 Pro is Google's flagship multimodal model supporting text, image, video, and audio inputs with a 1 million token context window.

Context 1.0M
Tier Flagship
Tools Supported
Modalities text, image, video, audio
Input from
$1.00 / 1M tokens
across 2 providers

API Pricing

Cheapest on Google Cloud 40% below avg
ProviderInput / 1MOutput / 1MSpeedTTFTUpdated
$1.00$6.00132 t/s24.4s4/13/2026
$2.00$12.00132 t/s24.4s4/14/2026
$2.00$12.00132 t/s24.4s4/13/2026

Prices updated daily. Last check: 4/14/2026

Model Details

General

Creator
Google
Family
Gemini
Tier
Flagship
Context Window
1.0M
Modalities
Text, Image, Video, Audio

Capabilities

Tool Calling
Yes
Open Source
No
Subtypes
Chat Completion, Code Generation
Aliases
gemini-3-1-pro-preview, gemini-3-1-pro

Strengths & Limitations

  • 1 million token context window enables processing of very long documents and conversations
  • Full multimodal support including text, image, video, and audio inputs
  • Tool calling capabilities with structured function execution
  • Code generation and programming assistance across multiple languages
  • Streaming response support for real-time interactions
  • Google's integration with search and knowledge systems
  • Handles both chat completion and code generation workloads
  • Proprietary model with no open-source weights available
  • Limited to Google's API ecosystem and approved providers
  • Benchmark performance data not publicly detailed
  • No information available on knowledge cutoff date

Key Features

1 million token context window
Multimodal input processing (text, image, video, audio)
Tool calling with function execution
Streaming responses
Code generation capabilities
Chat completion interface
JSON mode for structured outputs
Batch processing support

About Gemini 3.1 Pro

Gemini 3.1 Pro is Google's flagship model in the Gemini family, representing the company's most capable offering for complex AI tasks. As a proprietary model developed by Google DeepMind, it sits at the top tier of the Gemini lineup alongside other current frontier models like Claude Opus 4.6 and GPT-5.4. The model features comprehensive multimodal capabilities, processing text, images, video, and audio inputs within its exceptionally large 1 million token context window. This extensive context length allows for processing of very long documents, extended conversations, or large codebases in a single session. Gemini 3.1 Pro includes tool calling functionality and excels at both chat completion and code generation tasks. In practice, Gemini 3.1 Pro competes directly with other flagship models for demanding applications requiring sophisticated reasoning, complex problem-solving, and multimodal understanding. Its massive context window distinguishes it from many peers, making it particularly suitable for tasks involving extensive content analysis or long-form interactions.

Common Use Cases

Gemini 3.1 Pro is designed for complex, resource-intensive applications that require flagship-level capabilities. Its massive context window makes it ideal for analyzing lengthy documents, processing entire codebases, or maintaining context across very long conversations. The comprehensive multimodal support enables sophisticated applications involving document analysis with images, video content understanding, or audio processing. Organizations use it for advanced coding assistance, complex reasoning tasks, multimodal content creation, and AI agent development where the combination of large context and multimodal understanding provides significant advantages over smaller or text-only models.

Frequently Asked Questions

How much does Gemini 3.1 Pro cost per million tokens?

Gemini 3.1 Pro pricing varies by provider and pricing type (standard vs batch). Input and output tokens are typically priced differently. Check the pricing table above for current rates across all providers.

What is Gemini 3.1 Pro best used for?

Gemini 3.1 Pro excels at complex tasks requiring large context understanding and multimodal processing. This includes analyzing long documents, processing entire codebases, multimodal content creation, advanced reasoning tasks, and AI agent development where the 1 million token context window and comprehensive modality support provide clear advantages.

How does the 1 million token context window compare to other models?

Gemini 3.1 Pro's 1 million token context window is among the largest available in current flagship models, allowing it to process much longer inputs than models with smaller context windows. This enables use cases like processing entire books, large codebases, or maintaining context across very extended conversations that would exceed the limits of models with smaller context windows.