Google

Gemma 3 27B

Multimodal
Zero-eval
#1MMMU (val)
#1WMT24++
#1BIG-Bench Extra Hard
+5 more

by Google

About

Gemma 3 27B is a multimodal language model developed by Google. It achieves strong performance with an average score of 65.4% across 26 benchmarks. It excels particularly in GSM8k (95.9%), IFEval (90.4%), MATH (89.0%). The model shows particular specialization in math tasks with an average performance of 78.2%. It supports a 262K token context window for handling large documents. The model is available through 2 API providers. As a multimodal model, it can process and understand text, images, and other input formats seamlessly. It's licensed for commercial use, making it suitable for enterprise applications. Released in 2025, it represents Google's latest advancement in AI technology.

Pricing Range
Input (per 1M)$0.10 -$0.11
Output (per 1M)$0.20 -$0.20
Providers2
Timeline
AnnouncedMar 12, 2025
ReleasedMar 12, 2025
Specifications
Training Tokens14.0T
Capabilities
Multimodal
License & Family
License
Gemma
Benchmark Performance Overview
Performance metrics and category breakdown

Overall Performance

26 benchmarks
Average Score
65.4%
Best Score
95.9%
High Performers (80%+)
8

Performance Metrics

Max Context Window
262.1K
Avg Throughput
33.0 tok/s
Avg Latency
0ms

Top Categories

math
78.2%
factuality
74.9%
code
73.4%
vision
71.6%
general
53.5%
Benchmark Performance
Top benchmark scores with normalized values (0-100%)
Ranking Across Benchmarks
Position relative to other models on each benchmark

GSM8k

Rank #7 of 46
#4Claude 3.5 Sonnet
96.4%
#5Claude 3.5 Sonnet
96.4%
#6Llama 3.1 405B Instruct
96.8%
#7Gemma 3 27B
95.9%
#8Qwen2.5 32B Instruct
95.9%
#9Qwen2.5 72B Instruct
95.8%
#10DeepSeek-V2.5
95.1%

IFEval

Rank #5 of 37
#2Llama 3.3 70B Instruct
92.1%
#3Nova Pro
92.1%
#4Claude 3.7 Sonnet
93.2%
#5Gemma 3 27B
90.4%
#6Gemma 3 4B
90.2%
#7Kimi K2 Instruct
89.8%
#8Nova Lite
89.7%

MATH

Rank #4 of 63
#1Gemini 2.0 Flash
89.7%
#2o1
96.4%
#3o3-mini
97.9%
#4Gemma 3 27B
89.0%
#5Gemini 2.0 Flash-Lite
86.8%
#6Gemini 1.5 Pro
86.5%
#7o1-preview
85.5%

HumanEval

Rank #24 of 62
#21GPT-4.5
88.0%
#22Claude 3.5 Haiku
88.1%
#23o1
88.1%
#24Gemma 3 27B
87.8%
#25GPT-4o mini
87.2%
#26GPT-4 Turbo
87.1%
#27Qwen2.5 72B Instruct
86.6%

BIG-Bench Hard

Rank #4 of 21
#1Gemini 1.5 Pro
89.2%
#2Claude 3.5 Sonnet
93.1%
#3Claude 3.5 Sonnet
93.1%
#4Gemma 3 27B
87.6%
#5Claude 3 Opus
86.8%
#6Gemma 3 12B
85.7%
#7Gemini 1.5 Flash
85.5%
All Benchmark Results for Gemma 3 27B
Complete list of benchmark scores with detailed information
GSM8k
GSM8k benchmark
math
text
0.96
95.9%
Self-reported
IFEval
IFEval benchmark
code
text
0.90
90.4%
Self-reported
MATH
MATH benchmark
math
text
0.89
89.0%
Self-reported
HumanEval
HumanEval benchmark
code
text
0.88
87.8%
Self-reported
BIG-Bench Hard
BIG-Bench Hard benchmark
general
text
0.88
87.6%
Self-reported
DocVQA
DocVQA benchmark
vision
multimodal
0.87
86.6%
Self-reported
AI2D
AI2D benchmark
general
text
0.84
84.5%
Self-reported
Natural2Code
Natural2Code benchmark
code
text
0.84
84.5%
Self-reported
ChartQA
ChartQA benchmark
general
multimodal
0.78
78.0%
Self-reported
Global-MMLU-Lite
Global-MMLU-Lite benchmark
general
text
0.75
75.1%
Self-reported
Showing 1 to 10 of 26 benchmarks