Nemotron 3 Super
Nemotron 3 Super is a high-accuracy reasoning model designed for multi-agent systems, featuring a Hybrid Latent Mixture-of-Experts (MoE) Mamba-Transformer architecture. Announced by NVIDIA in December 2025 with availability expected in H1 2026, it is optimized for collaborative agent workflows and high-volume tasks like IT ticket automation. The model utilizes approximately 100 billion total parameters with 10 billion active per token, leveraging NVIDIA's NVFP4 training format and Multi-Token Prediction (MTP) for efficiency. It supports a 1 million token context window and is released under the NVIDIA Open Model License.
2025-12-15
100B (10B active)
Hybrid Latent Mixture-of-Experts (MoE) Mamba-Transformer
NVIDIA Open Model License
Specifications
- Parameters
- 100B (10B active)
- Architecture
- Hybrid Latent Mixture-of-Experts (MoE) Mamba-Transformer
- License
- NVIDIA Open Model License
- Context Window
- 1,000,000 tokens
- Training Data Cutoff
- null
- Type
- text
- Modalities
- text
Benchmark Scores
Advanced Specifications
- Model Family
- Nemotron 3
- Finetuned From
- null
- API Access
- Available
- Chat Interface
- Not Available
- Multilingual Support
- Yes
- Hardware Support
- NVIDIA BlackwellNVIDIA H100NVIDIA B200
Capabilities & Limitations
- Capabilities
- reasoningmulti-agent coordinationcodingtool useRAGlong context
- Notable Use Cases
- collaborative agentsIT ticket automationmulti-agent workflowslow-latency reasoning
- Function Calling Support
- Yes
- Tool Use Support
- Yes
Resources
- research.nvidia.com/labs/nemotron/Nemotron-3/
- nvidianews.nvidia.com/news/nvidia-debuts-nemotron-3-family-of-open-models
- developer.nvidia.com/blog/inside-nvidia-nemotron-3-techniques-tools-and-data-that-make-it-efficient-and-accurate/
- research.nvidia.com/labs/nemotron/files/NVIDIA-Nemotron-3-White-Paper.pdf