Qwen3.6-27B vs 35B, I prefer 35B but more people here post about 27B...

Reddit r/LocalLLaMA / 5/3/2026

💬 OpinionSignals & Early TrendsTools & Practical Usage

Key Points

  • The author reports getting better output quality and faster performance from Qwen3.6 35B than from 27B in their own local testing.
  • They suspect differences in how they run the models, noting that 27B results may be weaker for their multi-stage coding and internet-research pipelines even with simpler prompts and multi-step workflows.
  • The author uses OpenCode/Opencode in parts of their workflow and compares how well each model performs when applying the same prompting approaches.
  • They observe that 35B is frequently run with nvfp4 quant (or sometimes fp8) while 27B is run with fp8 or nvfp4 quant, suggesting the quantization/setup may affect outcomes.
  • They share their hardware setups (Mac Studio M4 Max 128GB RAM at home and work Mac M5 Max 48GB) to contextualize performance differences.

I've had better results quality wise with 35B AND it's much faster than 27B. Just curious cause I see lots of people post about 27B. Am I doing something wrong with 27B?

Use cases are multi-stage pipelines for coding and internet research. I also use Opencode a bit. All use cases I normally apply Opus to I've tried, as well as simpler prompts and mutli-step workflows. 35B seems to always perform as good or better and be much faster.

Edit:

35B is nvfp4 quant or sometimes fp8 and 27B is fp8 or nvfp4 quant

Edit 2:

I have 2 setups:

Home setup of Mac studio M4 Max 128Gb RAM, work mac M5 ~~ultra~~ max 48Gb ram.

submitted by /u/Snoo_27681
[link] [comments]