AIME 2024
general
text
About
AIME 2024 benchmark
Evaluation Stats
Total Models41
Organizations10
Verified Results0
Self-Reported41
Benchmark Details
Max Score1
Language
en
Performance Overview
Score distribution and top performers
Score Distribution
41 models
Top Score
95.8%
Average Score
72.8%
High Performers (80%+)
22Top Organizations
#1xAI
2 models
94.5%
#2Google
3 models
84.4%
#3IBM
2 models
81.2%
#4Anthropic
1 model
80.0%
#5Microsoft
2 models
78.3%
Leaderboard
Top 20 models ranked by performance
1
95.8%
Raw: 0.958
Self-reported
92.0%
Raw: 0.92
Self-reported
91.4%
Raw: 0.914
Self-reported
88.0%
Raw: 0.88
Self-reported
86.7%
Raw: 0.867
Self-reported
86.7%
Raw: 0.867
Self-reported
85.7%
Raw: 0.857
Self-reported
83.3%
Raw: 0.833
Self-reported
83.3%
Raw: 0.833
Self-reported
81.3%
Raw: 0.813
Self-reported
81.2%
Raw: 0.812
Self-reported
81.2%
Raw: 0.812
Self-reported
19
80.4%
Raw: 0.804
Self-reported
80.0%
Raw: 0.8
Self-reported
Showing top 20 of 41 models