
Qwen2.5-Coder 32B Instruct
Zero-eval
#1BigCodeBench-Full
#1BigCodeBench-Hard
#2MBPP
+1 more
by Alibaba
About
Qwen2.5-Coder 32B Instruct is a language model developed by Alibaba. It achieves strong performance with an average score of 64.9% across 15 benchmarks. It excels particularly in HumanEval (92.7%), GSM8k (91.1%), MBPP (90.2%). The model shows particular specialization in reasoning tasks with an average performance of 78.1%. It supports a 256K token context window for handling large documents. The model is available through 4 API providers. It's licensed for commercial use, making it suitable for enterprise applications. Released in 2024, it represents Alibaba's latest advancement in AI technology.
Pricing Range
Input (per 1M)$0.09 -$0.89
Output (per 1M)$0.09 -$0.89
Providers4
Timeline
AnnouncedSep 19, 2024
ReleasedSep 19, 2024
Specifications
Training Tokens5.5T
License & Family
License
Apache 2.0
Base ModelQwen2.5 32B Instruct
Benchmark Performance Overview
Performance metrics and category breakdown
Overall Performance
15 benchmarks
Average Score
64.9%
Best Score
92.7%
High Performers (80%+)
5Performance Metrics
Max Context Window
256.0KAvg Throughput
74.0 tok/sAvg Latency
0msTop Categories
reasoning
78.1%
math
74.2%
general
61.5%
code
58.2%
factuality
54.2%
Benchmark Performance
Top benchmark scores with normalized values (0-100%)
Ranking Across Benchmarks
Position relative to other models on each benchmark
HumanEval
Rank #4 of 62
#1Kimi K2 Instruct
93.3%
#2GPT-5
93.4%
#3Claude 3.5 Sonnet
93.7%
#4Qwen2.5-Coder 32B Instruct
92.7%
#5o1-mini
92.4%
#6Claude 3.5 Sonnet
92.0%
#7Mistral Large 2
92.0%
GSM8k
Rank #24 of 46
#21Qwen2 72B Instruct
91.1%
#22Llama 3.1 Nemotron 70B Instruct
91.4%
#23Qwen2.5 7B Instruct
91.6%
#24Qwen2.5-Coder 32B Instruct
91.1%
#25Gemini 1.5 Pro
90.8%
#26Grok-1.5
90.0%
#27Gemma 3 4B
89.2%
MBPP
Rank #2 of 31
#1Llama-3.3 Nemotron Super 49B v1
91.3%
#2Qwen2.5-Coder 32B Instruct
90.2%
#3Qwen2.5 72B Instruct
88.2%
#4Llama 3.1 Nemotron Nano 8B V1
84.6%
#5Qwen2.5 32B Instruct
84.0%
HellaSwag
Rank #14 of 24
#11Mistral NeMo Instruct
83.5%
#12Phi-3.5-MoE-instruct
83.8%
#13Qwen2.5 32B Instruct
85.2%
#14Qwen2.5-Coder 32B Instruct
83.0%
#15Gemma 2 9B
81.9%
#16Granite 3.3 8B Base
80.1%
#17Gemma 3n E4B Instructed LiteRT Preview
78.6%
Winogrande
Rank #8 of 19
#5Phi-3.5-MoE-instruct
81.3%
#6Qwen2.5 32B Instruct
82.0%
#7Gemma 2 27B
83.7%
#8Qwen2.5-Coder 32B Instruct
80.8%
#9Gemma 2 9B
80.6%
#10Mistral NeMo Instruct
76.8%
#11Ministral 8B Instruct
75.3%
All Benchmark Results for Qwen2.5-Coder 32B Instruct
Complete list of benchmark scores with detailed information
HumanEval HumanEval benchmark | code | text | 0.93 | 92.7% | Self-reported |
GSM8k GSM8k benchmark | math | text | 0.91 | 91.1% | Self-reported |
MBPP MBPP benchmark | code | text | 90.20 | 90.2% | Self-reported |
HellaSwag HellaSwag benchmark | reasoning | text | 0.83 | 83.0% | Self-reported |
Winogrande Winogrande benchmark | reasoning | text | 0.81 | 80.8% | Self-reported |
MMLU-Redux MMLU-Redux benchmark | general | text | 0.78 | 77.5% | Self-reported |
MMLU MMLU benchmark | general | text | 0.75 | 75.1% | Self-reported |
ARC-C ARC-C benchmark | reasoning | text | 0.70 | 70.5% | Self-reported |
MATH MATH benchmark | math | text | 0.57 | 57.2% | Self-reported |
TruthfulQA TruthfulQA benchmark | factuality | text | 0.54 | 54.2% | Self-reported |
Showing 1 to 10 of 15 benchmarks