AI Navigate

llama : add support for Nemotron 3 Super by danbev · Pull Request #20411 · ggml-org/llama.cpp

Reddit r/LocalLLaMA / 3/12/2026

📰 NewsDeveloper Stack & InfrastructureTools & Practical Usage

Key Points

  • A pull request to llama.cpp (PR #20411) adds support for NVIDIA Nemotron-3-Super-120B in GGML via GGUF.
  • The Nemotron-3-Super GGUF model artifact is hosted on HuggingFace at unsloth/NVIDIA-Nemotron-3-Super-120B-A12B-GGUF.
  • The PR was submitted by /u/jacek2023 and linked from Reddit’s LocalLLaMA discussion.
  • This update expands llama.cpp compatibility with large 120B parameter models, enabling easier experimentation and deployment.
  • The article provides direct links to the PR and the GGUF model page.