This is my first test with this model and Qwen impressed me. I will rate it 98% usable web os compared to my previous best 70% usable result from qwen3 next coder at q2.
Yes I know they train the models on these common prompts yet this is the best results I have seen even compared to a SOTA models.
~2100 lines of code used 38k context using opencode
Hardware: 24GB ddr5 + RTX4050
Quant: q4_k_xl
tg - 25 tk/s
llama-server \
--model /run/media/loq/New\ Volume/Models/unsloth/Qwen3.6-35B-A3B-GGUF/Qwen3.6-35B-A3B-UD-Q4_K_XL.gguf \
--port 1234 \
--host "0.0.0.0" \
--jinja \
-cmoe \
-t 8 -fa 1 -ctk q8_0 -ctv q8_0 \
--parallel 1 --fit-target 200 \
--temp 0.6 --top-p 0.95 --min-p 0.0 --top-k 20 --presence-penalty 0 --repeat-penalty 1.0
[link] [comments]




