AIME 2024

general
text
About

AIME 2024 benchmark

Evaluation Stats
Total Models41
Organizations10
Verified Results0
Self-Reported41
Benchmark Details
Max Score1
Language
en
Performance Overview
Score distribution and top performers

Score Distribution

41 models
Top Score
95.8%
Average Score
72.8%
High Performers (80%+)
22

Top Organizations

#1xAI
2 models
94.5%
#2Google
3 models
84.4%
#3IBM
2 models
81.2%
#4Anthropic
1 model
80.0%
#5Microsoft
2 models
78.3%
Leaderboard
Top 20 models ranked by performance
95.8%
Raw: 0.958
Self-reported
93.4%
Raw: 0.934
Self-reported
93.3%
Raw: 0.933
Self-reported
92.0%
Raw: 0.92
Self-reported
91.6%
Raw: 0.916
Self-reported
91.4%
Raw: 0.914
Self-reported
88.0%
Raw: 0.88
Self-reported
87.3%
Raw: 0.873
Self-reported
86.7%
Raw: 0.867
Self-reported
86.7%
Raw: 0.867
Self-reported
86.0%
Raw: 0.86
Self-reported
85.7%
Raw: 0.857
Self-reported
83.3%
Raw: 0.833
Self-reported
83.3%
Raw: 0.833
Self-reported
81.4%
Raw: 0.814
Self-reported
81.3%
Raw: 0.813
Self-reported
81.2%
Raw: 0.812
Self-reported
81.2%
Raw: 0.812
Self-reported
80.4%
Raw: 0.804
Self-reported
80.0%
Raw: 0.8
Self-reported
Showing top 20 of 41 models