
Llama 3.2 11B Instruct
Multimodal
Zero-eval
#1VQAv2 (test)
by Meta
About
Llama 3.2 11B Instruct is a multimodal language model developed by Meta. It achieves strong performance with an average score of 63.6% across 11 benchmarks. It excels particularly in AI2D (91.1%), DocVQA (88.4%), ChartQA (83.4%). It supports a 256K token context window for handling large documents. The model is available through 6 API providers. As a multimodal model, it can process and understand text, images, and other input formats seamlessly. Released in 2024, it represents Meta's latest advancement in AI technology.
Pricing Range
Input (per 1M)$0.05 -$0.20
Output (per 1M)$0.05 -$0.30
Providers6
Timeline
AnnouncedSep 25, 2024
ReleasedSep 25, 2024
Knowledge CutoffDec 31, 2023
Specifications
Capabilities
Multimodal
License & Family
License
Llama 3.2 Community License
Benchmark Performance Overview
Performance metrics and category breakdown
Overall Performance
11 benchmarks
Average Score
63.6%
Best Score
91.1%
High Performers (80%+)
3Performance Metrics
Max Context Window
256.0KAvg Throughput
116.8 tok/sAvg Latency
0msTop Categories
general
70.1%
vision
61.8%
math
57.4%
Benchmark Performance
Top benchmark scores with normalized values (0-100%)
Ranking Across Benchmarks
Position relative to other models on each benchmark
AI2D
Rank #6 of 17
#3Llama 3.2 90B Instruct
92.3%
#4Mistral Small 3.2 24B Instruct
92.9%
#5Pixtral Large
93.8%
#6Llama 3.2 11B Instruct
91.1%
#7Qwen2.5 VL 72B Instruct
88.4%
#8Grok-1.5V
88.3%
#9Gemma 3 27B
84.5%
DocVQA
Rank #21 of 26
#18DeepSeek VL2 Tiny
88.9%
#19Llama 3.2 90B Instruct
90.1%
#20Pixtral-12B
90.7%
#21Llama 3.2 11B Instruct
88.4%
#22Gemma 3 12B
87.1%
#23Gemma 3 27B
86.6%
#24Grok-1.5V
85.6%
ChartQA
Rank #16 of 24
#13DeepSeek VL2 Small
84.5%
#14Qwen2.5-Omni-7B
85.3%
#15Llama 3.2 90B Instruct
85.5%
#16Llama 3.2 11B Instruct
83.4%
#17Pixtral-12B
81.8%
#18Phi-3.5-vision-instruct
81.8%
#19Phi-4-multimodal-instruct
81.4%
VQAv2 (test)
Rank #1 of 1
#1Llama 3.2 11B Instruct
75.2%
MMLU
Rank #57 of 78
#54Qwen2.5-Coder 32B Instruct
75.1%
#55Claude 3 Haiku
75.2%
#56Gemma 2 27B
75.2%
#57Llama 3.2 11B Instruct
73.0%
#58Gemini 1.0 Pro
71.8%
#59Gemma 2 9B
71.3%
#60Qwen2 7B Instruct
70.5%
All Benchmark Results for Llama 3.2 11B Instruct
Complete list of benchmark scores with detailed information
AI2D AI2D benchmark | general | text | 0.91 | 91.1% | Self-reported |
DocVQA DocVQA benchmark | vision | multimodal | 0.88 | 88.4% | Self-reported |
ChartQA ChartQA benchmark | general | multimodal | 0.83 | 83.4% | Self-reported |
VQAv2 (test) VQAv2 (test) benchmark | vision | multimodal | 0.75 | 75.2% | Self-reported |
MMLU MMLU benchmark | general | text | 0.73 | 73.0% | Self-reported |
MGSM MGSM benchmark | math | text | 0.69 | 68.9% | Self-reported |
MATH MATH benchmark | math | text | 0.52 | 51.9% | Self-reported |
MathVista MathVista benchmark | math | text | 0.52 | 51.5% | Self-reported |
MMMU MMMU benchmark | vision | multimodal | 0.51 | 50.7% | Self-reported |
MMMU-Pro MMMU-Pro benchmark | vision | multimodal | 0.33 | 33.0% | Self-reported |
Showing 1 to 10 of 11 benchmarks