NVIDIA has released Nemotron 3 Super, an open hybrid Mamba Transformer MoE model designed for agentic reasoning tasks.
The model features 120 billion parameters in a Mixture of Experts (MoE) architecture, with 12 billion parameters active at any given time.
Nemotron 3 Super aims to improve efficiency and performance in large-scale AI models by activating only a subset of the total parameters during inference.
The release is detailed on NVIDIA's developer blog, highlighting its technical innovations and potential applications in AI agentic reasoning.