| I just used the new Apex Quantization on QWEN Coder 80B Created an Important Matrix using Code examples This should be the fastest best at coding 80B Next Coder around It's what I'm using for STACKS! so I thought I would share with the community It's insanely fast and the size has been shrunk down to 54.1GB https://huggingface.co/stacksnathan/Qwen3-Coder-Next-80B-APEX-I-Quality-GGUF [link] [comments] |
Fastest QWEN Coder 80B Next
Reddit r/LocalLLaMA / 4/5/2026
💬 OpinionSignals & Early TrendsTools & Practical UsageModels & Research
Key Points
- A user reports they used “Apex Quantization” with Qwen Coder 80B (Qwen3-Coder-Next-80B) and claims it delivers very high coding performance.
- The post states the quantized model size was reduced to about 54.1GB while maintaining strong coding results.
- The user shares a Hugging Face release: stacksnathan/Qwen3-Coder-Next-80B-APEX-I-Quality-GGUF.
- They mention creating/using an “Important Matrix” based on code examples and intend to use this setup for “STACKS,” encouraging others to try it.




