OpenAI

GPT-4.1 nano

Multimodal
Zero-eval
#2OpenAI-MRCR: 2 needle 1M
#3Graphwalks parents >128k
#3Graphwalks BFS >128k

by OpenAI

About

GPT-4.1 nano is a multimodal language model developed by OpenAI. The model shows competitive results across 25 benchmarks. It excels particularly in MMLU (80.1%), IFEval (74.5%), CharXiv-D (73.9%). With a 1.1M token context window, it can handle extensive documents and complex multi-turn conversations. The model is available through 1 API provider. As a multimodal model, it can process and understand text, images, and other input formats seamlessly. Released in 2025, it represents OpenAI's latest advancement in AI technology.

Pricing Range
Input (per 1M)$0.10 -$0.10
Output (per 1M)$0.40 -$0.40
Providers1
Timeline
AnnouncedApr 14, 2025
ReleasedApr 14, 2025
Knowledge CutoffMay 31, 2024
Specifications
Capabilities
Multimodal
License & Family
License
Proprietary
Benchmark Performance Overview
Performance metrics and category breakdown

Overall Performance

25 benchmarks
Average Score
36.1%
Best Score
80.1%
High Performers (80%+)
1

Performance Metrics

Max Context Window
1.1M
Avg Throughput
200.0 tok/s
Avg Latency
2ms

Top Categories

code
74.5%
math
56.2%
vision
55.4%
long_context
48.3%
general
32.4%
Benchmark Performance
Top benchmark scores with normalized values (0-100%)
Ranking Across Benchmarks
Position relative to other models on each benchmark

MMLU

Rank #45 of 78
#42Llama 3.1 Nemotron 70B Instruct
80.2%
#43DeepSeek-V2.5
80.4%
#44Nova Lite
80.5%
#45GPT-4.1 nano
80.1%
#46Qwen2.5 14B Instruct
79.7%
#47Llama 4 Scout
79.6%
#48Claude 3 Sonnet
79.0%

IFEval

Rank #33 of 37
#30Granite 3.3 8B Base
74.8%
#31Granite 3.3 8B Instruct
74.8%
#32Llama 3.2 3B Instruct
77.4%
#33GPT-4.1 nano
74.5%
#34Qwen2.5 7B Instruct
71.2%
#35IBM Granite 4.0 Tiny Preview
63.0%
#36Phi 4
63.0%

CharXiv-D

Rank #5 of 5
#2GPT-4o
85.3%
#3GPT-4.1
87.9%
#4GPT-4.1 mini
88.4%
#5GPT-4.1 nano
73.9%

MMMLU

Rank #12 of 13
#9Phi-3.5-MoE-instruct
69.9%
#10GPT-4.1 mini
78.5%
#11GPT-4o
81.4%
#12GPT-4.1 nano
66.9%
#13Phi-3.5-mini-instruct
55.4%

OpenAI-MRCR: 2 needle 1M

Rank #2 of 3
#1GPT-4.1
61.2%
#2GPT-4.1 nano
59.9%
#3GPT-4.1 mini
23.5%
All Benchmark Results for GPT-4.1 nano
Complete list of benchmark scores with detailed information
MMLU
MMLU benchmark
general
text
0.80
80.1%
Self-reported
IFEval
IFEval benchmark
code
text
0.74
74.5%
Self-reported
CharXiv-D
CharXiv-D benchmark
general
text
0.74
73.9%
Self-reported
MMMLU
MMMLU benchmark
general
text
0.67
66.9%
Self-reported
OpenAI-MRCR: 2 needle 1M
OpenAI-MRCR: 2 needle 1M benchmark
long_context
text
0.60
59.9%
Self-reported
Multi-IF
Multi-IF benchmark
general
text
0.57
57.2%
Self-reported
MathVista
MathVista benchmark
math
text
0.56
56.2%
Self-reported
MMMU
MMMU benchmark
vision
multimodal
0.55
55.4%
Self-reported
GPQA
GPQA benchmark
general
text
0.50
50.3%
Self-reported
COLLIE
COLLIE benchmark
general
text
0.42
42.5%
Self-reported
Showing 1 to 10 of 25 benchmarks