🚀 Website under development • Launching soon

GLM-4.5-Air

Zero-eval
#1TAU-bench Airline
#2MATH-500
#2BFCL-v3
+3 more

by Zhipu AI

About

GLM-4.5-Air is a language model developed by Zhipu AI. It achieves strong performance with an average score of 60.8% across 14 benchmarks. It excels particularly in MATH-500 (98.1%), AIME 2024 (89.4%), MMLU-Pro (81.4%). It's licensed for commercial use, making it suitable for enterprise applications. Released in 2025, it represents Zhipu AI's latest advancement in AI technology.

Timeline
AnnouncedJul 28, 2025
ReleasedJul 28, 2025
Specifications
License & Family
License
MIT
Benchmark Performance Overview
Performance metrics and category breakdown

Overall Performance

14 benchmarks
Average Score
60.8%
Best Score
98.1%
High Performers (80%+)
3

Top Categories

math
98.1%
general
76.0%
agents
53.3%
code
46.0%
reasoning
37.7%
Additional Information

Content coming soon...

All Benchmark Results for GLM-4.5-Air
Complete list of benchmark scores with detailed information
MATH-500
math
text
0.98
98.1%
Self-reported
AIME 2024
general
text
0.89
89.4%
Self-reported
MMLU-Pro
general
text
0.81
81.4%
Self-reported
TAU-bench Retail
agents
text
0.78
77.9%
Self-reported
BFCL-v3
general
text
0.76
76.4%
Self-reported
GPQA
general
text
0.75
75.0%
Self-reported
LiveCodeBench
code
text
0.71
70.7%
Self-reported
AA-Index
reasoning
text
0.65
64.8%
Self-reported
TAU-bench Airline
agents
text
0.61
60.8%
Self-reported
SWE-Bench Verified
general
text
0.58
57.6%
Self-reported
Showing 1 to 10 of 14 benchmarks