Google

Gemini 1.0 Pro

Zero-eval
#1BIG-Bench

by Google

About

Gemini 1.0 Pro is a language model developed by Google. The model shows competitive results across 9 benchmarks. Notable strengths include BIG-Bench (75.0%), MMLU (71.8%), WMT23 (71.7%). The model is available through 1 API provider. Released in 2024, it represents Google's latest advancement in AI technology.

Pricing Range
Input (per 1M)$0.50 -$0.50
Output (per 1M)$1.50 -$1.50
Providers1
Timeline
AnnouncedFeb 15, 2024
ReleasedFeb 15, 2024
Knowledge CutoffFeb 1, 2024
Specifications
License & Family
License
Proprietary
Benchmark Performance Overview
Performance metrics and category breakdown

Overall Performance

9 benchmarks
Average Score
48.4%
Best Score
75.0%
High Performers (80%+)
0

Performance Metrics

Max Context Window
41.0K
Avg Throughput
120.0 tok/s
Avg Latency
0ms

Top Categories

general
51.4%
vision
47.9%
math
39.6%
Benchmark Performance
Top benchmark scores with normalized values (0-100%)
Ranking Across Benchmarks
Position relative to other models on each benchmark

BIG-Bench

Rank #1 of 3
#1Gemini 1.0 Pro
75.0%
#2Gemma 2 27B
74.9%
#3Gemma 2 9B
68.2%

MMLU

Rank #58 of 78
#55Llama 3.2 11B Instruct
73.0%
#56Qwen2.5-Coder 32B Instruct
75.1%
#57Claude 3 Haiku
75.2%
#58Gemini 1.0 Pro
71.8%
#59Gemma 2 9B
71.3%
#60Qwen2 7B Instruct
70.5%
#61GPT-3.5 Turbo
69.8%

WMT23

Rank #4 of 4
#1Gemini 1.5 Flash 8B
72.6%
#2Gemini 1.5 Flash
74.1%
#3Gemini 1.5 Pro
75.1%
#4Gemini 1.0 Pro
71.7%

EgoSchema

Rank #9 of 9
#6Gemini 2.0 Flash-Lite
67.2%
#7Qwen2.5-Omni-7B
68.6%
#8Nova Lite
71.4%
#9Gemini 1.0 Pro
55.7%

MMMU

Rank #49 of 52
#46DeepSeek VL2 Small
48.0%
#47Llama 3.2 11B Instruct
50.7%
#48DeepSeek VL2
51.1%
#49Gemini 1.0 Pro
47.9%
#50Phi-3.5-vision-instruct
43.0%
#51DeepSeek VL2 Tiny
40.7%
#52GPT-3.5 Turbo
0.0%
All Benchmark Results for Gemini 1.0 Pro
Complete list of benchmark scores with detailed information
BIG-Bench
BIG-Bench benchmark
general
text
0.75
75.0%
Unverified
MMLU
MMLU benchmark
general
text
0.72
71.8%
Self-reported
WMT23
WMT23 benchmark
general
text
0.72
71.7%
Unverified
EgoSchema
EgoSchema benchmark
general
text
0.56
55.7%
Self-reported
MMMU
MMMU benchmark
vision
multimodal
0.48
47.9%
Unverified
MathVista
MathVista benchmark
math
text
0.47
46.6%
Unverified
MATH
MATH benchmark
math
text
0.33
32.6%
Unverified
GPQA
GPQA benchmark
general
text
0.28
27.9%
Unverified
FLEURS
FLEURS benchmark
general
text
6.40
6.4%
Unverified