NVIDIA logo

Nemotron 3 Super

NVIDIAOpen WeightsPending Human Review

Nemotron 3 Super is a high-accuracy reasoning model designed for multi-agent systems, featuring a Hybrid Latent Mixture-of-Experts (MoE) Mamba-Transformer architecture. Announced by NVIDIA in December 2025 with availability expected in H1 2026, it is optimized for collaborative agent workflows and high-volume tasks like IT ticket automation. The model utilizes approximately 100 billion total parameters with 10 billion active per token, leveraging NVIDIA's NVFP4 training format and Multi-Token Prediction (MTP) for efficiency. It supports a 1 million token context window and is released under the NVIDIA Open Model License.

2025-12-15
100B (10B active)
Hybrid Latent Mixture-of-Experts (MoE) Mamba-Transformer
NVIDIA Open Model License

Specifications

Parameters
100B (10B active)
Architecture
Hybrid Latent Mixture-of-Experts (MoE) Mamba-Transformer
License
NVIDIA Open Model License
Context Window
1,000,000 tokens
0
Training Data Cutoff
null
Type
text
Modalities
text

Benchmark Scores

Advanced Specifications

Model Family
Nemotron 3
Finetuned From
null
API Access
Available
Chat Interface
Not Available
Multilingual Support
Yes
Hardware Support
NVIDIA BlackwellNVIDIA H100NVIDIA B200

Capabilities & Limitations

Capabilities
reasoningmulti-agent coordinationcodingtool useRAGlong context
Notable Use Cases
collaborative agentsIT ticket automationmulti-agent workflowslow-latency reasoning
Function Calling Support
Yes
Tool Use Support
Yes

Related Models