Google pricing
Gemini 3 Flash Cost Calculator
Gemini 3 Flash balances speed and capability with 1M context. Native audio + vision. $0.05/1M cached input makes it dominant for high-volume multimodal pipelines.
Input
$0.50
per 1M tokens
Output
$3.00
per 1M tokens
Context window
1049K
tokens
Released
2025-12
Cutoff 2025-10
≈ Estimated tokenizer·$0.50 in·$3.00 out (per 1M)
Quick start with a use case
Total cost per call$0.002000
Input$0.000500
Output$0.001500
Cost comparison
Standard
$0.002000
With Caching
$0.001775
Save 11% ↓
With Batch
$0.001000
Save 50% ↓
Detailed pricing
Gemini 3 Flash pricing breakdown
All pricing dimensions including caching and batch discounts.
| Type | Price (per 1M tokens) |
|---|---|
| Input | $0.5000 |
| Output | $3.0000 |
| Cached input | $0.0500 |
| Batch input | $0.2500 |
| Batch output | $1.5000 |
Last verified 2026-05-12 · Google official pricing · ⚠️ Spotted a wrong price? Report in 30s →
How it compares
Gemini 3 Flash vs alternatives
Single-call cost (1000 input + 500 output tokens) ranked from cheapest.
| Model | Per call |
|---|---|
Gemini 3 Flash Google · this page | $0.002000 |
| DeepSeek V3.2 DeepSeek | $0.000480 |
| GPT-5 mini OpenAI | $0.001250 |
| Mistral Large 3 Mistral | $0.001250 |
| o4-mini OpenAI | $0.003300 |
| Claude Haiku 4.5 Anthropic | $0.003500 |
Recommended use
When to choose Gemini 3 Flash
Gemini 3 Flash shines for general-purpose tasks, image and document understanding, high-throughput, low-latency tasks, and audio understanding and generation. Token counts are estimated within ~10-20% margin.
✓Context window of 1049K tokens handles entire codebases or book-length documents.
✓Prompt caching available — significant savings for repeated system prompts.
✓Batch API support for non-realtime workloads at ~50% discount.
✓Tool / function calling supported.
FAQ
Frequently asked questions
Gemini 3 Flash costs $0.50 per 1M input tokens and $3.00 per 1M output tokens. A typical chat call (1000 input + 500 output tokens) costs approximately $0.0020. Use the calculator above to estimate your specific use case.
Related
Other models you might consider
DeepSeek
DeepSeek V3.2
DeepSeek's open-weight MoE — top cost/quality ratio
$0.28 in$0.40 out
OpenAI
GPT-5 mini
OpenAI's affordable small model with 272K context
$0.25 in$2.00 out
Mistral
Mistral Large 3
Mistral's flagship multilingual model — EU-friendly
$0.50 in$1.50 out
OpenAI
o4-mini
OpenAI's reasoning model — sharp at STEM/coding
$1.10 in$4.40 out
Anthropic
Claude Haiku 4.5
Anthropic's fast affordable workhorse
$1.00 in$5.00 out