NVIDIA Unveils Nemotron 3: Nano, Super, and Ultra Open Models for Agentic AI
NVIDIA debuted the Nemotron 3 family at GTC 2026, releasing Nano immediately and previewing Super (49B) and Ultra (253B) variants, along with three trillion tokens of pre-training data.
NVIDIA announced the Nemotron 3 family of open models at GTC 2026, positioning the lineup as the most efficient open models for building agentic AI applications. The family spans three tiers — Nano, Super, and Ultra — each tuned for different deployment contexts, from edge devices and multi-agent swarms to high-accuracy enterprise workloads running on multi-GPU clusters.
The Nemotron 3 Nano model launched immediately and delivers a headline benchmark: four times higher throughput than its Nemotron 2 predecessor, achieved through a hybrid mixture-of-experts architecture that dramatically reduces the compute required per token. For developers building multi-agent systems at scale, where thousands of inference calls may happen in parallel, this throughput improvement directly translates to lower latency and cost at runtime.
Nemotron 3 Super, at 49 billion parameters, and Nemotron 3 Ultra, at 253 billion parameters, are targeted for availability in the first half of 2026. Both are designed for applications requiring high reasoning accuracy — complex coding tasks, long-document analysis, and autonomous agent pipelines that need reliable, step-by-step problem decomposition. NVIDIA says the Super and Ultra models match or outperform much larger models from other providers on standard reasoning benchmarks.
In a notable move for the open-source community, NVIDIA is releasing not just the model weights but also the data used to train them: three trillion tokens of pre-training data and 18 million samples of post-training instruction data. This level of transparency is rare among frontier model releases and gives researchers and fine-tuners the ability to understand, audit, and extend the models in ways that are impossible with proprietary systems.
The release complements NVIDIA's broader push into physical AI and robotics, announced at the same GTC event alongside the GR00T N1.7 open model for humanoid robot training. Together, the announcements signal that NVIDIA sees open model releases as a strategic tool for expanding the AI software ecosystem around its hardware — a long-term bet that the more developers build with NVIDIA-origin models, the more they deploy on NVIDIA chips.