| submitted by /u/jacek2023 [link] [comments] |
llama : add support for Nemotron 3 Super by danbev · Pull Request #20411 · ggml-org/llama.cpp
Reddit r/LocalLLaMA / 3/12/2026
📰 NewsDeveloper Stack & InfrastructureTools & Practical Usage
Key Points
- A pull request to llama.cpp (PR #20411) adds support for NVIDIA Nemotron-3-Super-120B in GGML via GGUF.
- The Nemotron-3-Super GGUF model artifact is hosted on HuggingFace at unsloth/NVIDIA-Nemotron-3-Super-120B-A12B-GGUF.
- The PR was submitted by /u/jacek2023 and linked from Reddit’s LocalLLaMA discussion.
- This update expands llama.cpp compatibility with large 120B parameter models, enabling easier experimentation and deployment.
- The article provides direct links to the PR and the GGUF model page.
Related Articles
I Was Wrong About AI Coding Assistants. Here's What Changed My Mind (and What I Built About It).
Dev.to
Qwen3.5-122B-A10B Uncensored (Aggressive) — GGUF Release + new K_P Quants
Reddit r/LocalLLaMA
Die besten AI Tools fuer Digital Nomads 2026
Dev.to
I Built the Most Feature-Complete MCP Server for Obsidian — Here's How
Dev.to
A supervisor or "manager" Al agent is the wrong way to control Al
Reddit r/artificial