AI Navigate

Nemotron 3 Super Released

Reddit r/LocalLLaMA / 3/12/2026

📰 NewsModels & Research

Key Points

  • NVIDIA has released Nemotron 3 Super, an open hybrid Mamba Transformer MoE model designed for agentic reasoning tasks.
  • The model features 120 billion parameters in a Mixture of Experts (MoE) architecture, with 12 billion parameters active at any given time.
  • Nemotron 3 Super aims to improve efficiency and performance in large-scale AI models by activating only a subset of the total parameters during inference.
  • The release is detailed on NVIDIA's developer blog, highlighting its technical innovations and potential applications in AI agentic reasoning.