Meta

Llama 3.2 11B Instruct

Multimodal
Zero-eval
#1VQAv2 (test)

by Meta

About

Llama 3.2 11B Instruct is a multimodal language model developed by Meta. It achieves strong performance with an average score of 63.6% across 11 benchmarks. It excels particularly in AI2D (91.1%), DocVQA (88.4%), ChartQA (83.4%). It supports a 256K token context window for handling large documents. The model is available through 6 API providers. As a multimodal model, it can process and understand text, images, and other input formats seamlessly. Released in 2024, it represents Meta's latest advancement in AI technology.

Pricing Range
Input (per 1M)$0.05 -$0.20
Output (per 1M)$0.05 -$0.30
Providers6
Timeline
AnnouncedSep 25, 2024
ReleasedSep 25, 2024
Knowledge CutoffDec 31, 2023
Specifications
Capabilities
Multimodal
License & Family
License
Llama 3.2 Community License
Benchmark Performance Overview
Performance metrics and category breakdown

Overall Performance

11 benchmarks
Average Score
63.6%
Best Score
91.1%
High Performers (80%+)
3

Performance Metrics

Max Context Window
256.0K
Avg Throughput
116.8 tok/s
Avg Latency
0ms

Top Categories

general
70.1%
vision
61.8%
math
57.4%
Benchmark Performance
Top benchmark scores with normalized values (0-100%)
Ranking Across Benchmarks
Position relative to other models on each benchmark

AI2D

Rank #6 of 17
#3Llama 3.2 90B Instruct
92.3%
#4Mistral Small 3.2 24B Instruct
92.9%
#5Pixtral Large
93.8%
#6Llama 3.2 11B Instruct
91.1%
#7Qwen2.5 VL 72B Instruct
88.4%
#8Grok-1.5V
88.3%
#9Gemma 3 27B
84.5%

DocVQA

Rank #21 of 26
#18DeepSeek VL2 Tiny
88.9%
#19Llama 3.2 90B Instruct
90.1%
#20Pixtral-12B
90.7%
#21Llama 3.2 11B Instruct
88.4%
#22Gemma 3 12B
87.1%
#23Gemma 3 27B
86.6%
#24Grok-1.5V
85.6%

ChartQA

Rank #16 of 24
#13DeepSeek VL2 Small
84.5%
#14Qwen2.5-Omni-7B
85.3%
#15Llama 3.2 90B Instruct
85.5%
#16Llama 3.2 11B Instruct
83.4%
#17Pixtral-12B
81.8%
#18Phi-3.5-vision-instruct
81.8%
#19Phi-4-multimodal-instruct
81.4%

VQAv2 (test)

Rank #1 of 1
#1Llama 3.2 11B Instruct
75.2%

MMLU

Rank #57 of 78
#54Qwen2.5-Coder 32B Instruct
75.1%
#55Claude 3 Haiku
75.2%
#56Gemma 2 27B
75.2%
#57Llama 3.2 11B Instruct
73.0%
#58Gemini 1.0 Pro
71.8%
#59Gemma 2 9B
71.3%
#60Qwen2 7B Instruct
70.5%
All Benchmark Results for Llama 3.2 11B Instruct
Complete list of benchmark scores with detailed information
AI2D
AI2D benchmark
general
text
0.91
91.1%
Self-reported
DocVQA
DocVQA benchmark
vision
multimodal
0.88
88.4%
Self-reported
ChartQA
ChartQA benchmark
general
multimodal
0.83
83.4%
Self-reported
VQAv2 (test)
VQAv2 (test) benchmark
vision
multimodal
0.75
75.2%
Self-reported
MMLU
MMLU benchmark
general
text
0.73
73.0%
Self-reported
MGSM
MGSM benchmark
math
text
0.69
68.9%
Self-reported
MATH
MATH benchmark
math
text
0.52
51.9%
Self-reported
MathVista
MathVista benchmark
math
text
0.52
51.5%
Self-reported
MMMU
MMMU benchmark
vision
multimodal
0.51
50.7%
Self-reported
MMMU-Pro
MMMU-Pro benchmark
vision
multimodal
0.33
33.0%
Self-reported
Showing 1 to 10 of 11 benchmarks