| submitted by /u/jacek2023 [link] [comments] |
llama : add support for Nemotron 3 Super by danbev · Pull Request #20411 · ggml-org/llama.cpp
Reddit r/LocalLLaMA / 3/12/2026
📰 NewsDeveloper Stack & InfrastructureTools & Practical Usage
Key Points
- A pull request to llama.cpp (PR #20411) adds support for NVIDIA Nemotron-3-Super-120B in GGML via GGUF.
- The Nemotron-3-Super GGUF model artifact is hosted on HuggingFace at unsloth/NVIDIA-Nemotron-3-Super-120B-A12B-GGUF.
- The PR was submitted by /u/jacek2023 and linked from Reddit’s LocalLLaMA discussion.
- This update expands llama.cpp compatibility with large 120B parameter models, enabling easier experimentation and deployment.
- The article provides direct links to the PR and the GGUF model page.




