About

GPU and AI company

Portfolio Stats
Total Models4
Multimodal0
Benchmarks Run31
Avg Performance75.0%
Latest Release
Release Timeline
Recent model releases by year
2025
3 models
2024
1 model
Performance Overview
Top models and benchmark performance

Benchmark Categories

math
5
92.5%
code
6
80.8%
reasoning
3
79.8%
general
13
66.7%
roleplay
3
60.6%

Model Statistics

Multimodal Ratio
0%
Models with Providers
0

All Models

Complete portfolio of 4 models with advanced filtering

LicenseLinks
NVIDIALlama 3.1 Nemotron Ultra 253B v1
A 253B parameter derivative of Meta Llama 3.1 405B Instruct, developed by NVIDIA using Neural Architecture Search (NAS) and vertical compression. It underwent multi-phase post-training (SFT for Math, Code, Reasoning, Chat, Tool Calling; RL with GRPO) to enhance reasoning and instruction-following. Optimized for accuracy/efficiency tradeoff on NVIDIA GPUs. Supports 128k context.
Apr 7, 2025
Llama 3.1 Community License
---66.3%-
NVIDIALlama-3.3 Nemotron Super 49B v1
Llama-3.3-Nemotron-Super-49B-v1 is a large language model (LLM) derived from Meta Llama-3.3-70B-Instruct. It's post-trained for reasoning, chat, RAG, and tool calling, offering a balance between accuracy and efficiency (optimized for single H100). It underwent multi-phase post-training including SFT and RL (RLOO, RPO).
Mar 18, 2025
Llama 3.1 Community License
----91.3%
NVIDIALlama 3.1 Nemotron Nano 8B V1
Llama-3.1-Nemotron-Nano-8B-v1 is a large language model (LLM) which is a derivative of Meta Llama-3.1-8B-Instruct (AKA the reference model). It is a reasoning model that is post trained for reasoning, human chat preferences, and tasks, such as RAG and tool calling.
Mar 18, 2025
Llama 3.1 Community License
----84.6%
NVIDIALlama 3.1 Nemotron 70B Instruct
A large language model customized by NVIDIA to improve the helpfulness of LLM generated responses. It is a fine-tuned version of Llama 3.1 70B Instruct. The model was trained using RLHF (REINFORCE) with HelpSteer2-Preference prompts.
Oct 1, 2024
Llama 3.1 Community License
-----