Google

Gemini 2.0 Flash

Multimodal
Zero-eval
#1Natural2Code
#1HiddenMath
#1CoVoST2
+2 more

by Google

About

Gemini 2.0 Flash is a multimodal language model developed by Google. It achieves strong performance with an average score of 66.7% across 13 benchmarks. It excels particularly in Natural2Code (92.9%), MATH (89.7%), FACTS Grounding (83.6%). With a 1.1M token context window, it can handle extensive documents and complex multi-turn conversations. The model is available through 1 API provider. As a multimodal model, it can process and understand text, images, and other input formats seamlessly. Released in 2024, it represents Google's latest advancement in AI technology.

Pricing Range
Input (per 1M)$0.10 -$0.10
Output (per 1M)$0.40 -$0.40
Providers1
Timeline
AnnouncedDec 1, 2024
ReleasedDec 1, 2024
Knowledge CutoffAug 1, 2024
Specifications
Capabilities
Multimodal
License & Family
License
Proprietary
Benchmark Performance Overview
Performance metrics and category breakdown

Overall Performance

13 benchmarks
Average Score
66.7%
Best Score
92.9%
High Performers (80%+)
3

Performance Metrics

Max Context Window
1.1M
Avg Throughput
183.0 tok/s
Avg Latency
0ms

Top Categories

factuality
83.6%
math
76.4%
vision
70.7%
general
62.6%
code
61.4%
Benchmark Performance
Top benchmark scores with normalized values (0-100%)
Ranking Across Benchmarks
Position relative to other models on each benchmark

Natural2Code

Rank #1 of 8
#1Gemini 2.0 Flash
92.9%
#2Gemini 1.5 Pro
85.4%
#3Gemma 3 27B
84.5%
#4Gemma 3 12B
80.7%

MATH

Rank #3 of 63
#1o1
96.4%
#2o3-mini
97.9%
#3Gemini 2.0 Flash
89.7%
#4Gemma 3 27B
89.0%
#5Gemini 2.0 Flash-Lite
86.8%
#6Gemini 1.5 Pro
86.5%

FACTS Grounding

Rank #4 of 9
#1Gemini 2.5 Flash-Lite
84.1%
#2Gemini 2.5 Flash
85.3%
#3Gemini 2.5 Pro Preview 06-05
87.8%
#4Gemini 2.0 Flash
83.6%
#5Gemini 2.0 Flash-Lite
83.6%
#6Gemma 3 12B
75.8%
#7Gemma 3 27B
74.9%

MMLU-Pro

Rank #8 of 60
#5Claude 3.5 Sonnet
77.6%
#6Llama 4 Maverick
80.5%
#7Kimi K2 Instruct
81.1%
#8Gemini 2.0 Flash
76.4%
#9Claude 3.5 Sonnet
76.1%
#10Phi 4 Reasoning Plus
76.0%
#11DeepSeek-V3
75.9%

EgoSchema

Rank #5 of 9
#2Nova Pro
72.1%
#3GPT-4o
72.2%
#4Qwen2.5 VL 72B Instruct
76.2%
#5Gemini 2.0 Flash
71.5%
#6Nova Lite
71.4%
#7Qwen2.5-Omni-7B
68.6%
#8Gemini 2.0 Flash-Lite
67.2%
All Benchmark Results for Gemini 2.0 Flash
Complete list of benchmark scores with detailed information
Natural2Code
Natural2Code benchmark
code
text
0.93
92.9%
Self-reported
MATH
MATH benchmark
math
text
0.90
89.7%
Self-reported
FACTS Grounding
FACTS Grounding benchmark
factuality
text
0.84
83.6%
Self-reported
MMLU-Pro
MMLU-Pro benchmark
general
text
0.76
76.4%
Self-reported
EgoSchema
EgoSchema benchmark
general
text
0.71
71.5%
Self-reported
MMMU
MMMU benchmark
vision
multimodal
0.71
70.7%
Self-reported
MRCR
MRCR benchmark
general
text
0.69
69.2%
Self-reported
HiddenMath
HiddenMath benchmark
math
text
0.63
63.0%
Self-reported
GPQA
GPQA benchmark
general
text
0.62
62.1%
Self-reported
Bird-SQL (dev)
Bird-SQL (dev) benchmark
general
text
0.57
56.9%
Self-reported
Showing 1 to 10 of 13 benchmarks