I'm running qwen3.6-35b-a3b with 8 bit quant and 64k context thru OpenCode on my mbp m5 max 128gb and it's as good as claude

Reddit r/LocalLLaMA / 4/19/2026

💬 OpinionDeveloper Stack & InfrastructureSignals & Early TrendsTools & Practical Usage

Key Points

  • The author reports testing several local language models and trying the newly available Qwen 3.6 model in LM Studio.
  • They claim Qwen3.6-35B with 8-bit quantization and a 64K context window runs very fast on an M5 Max MacBook Pro with 128GB RAM.
  • The model is described as performing well on long research-style tasks, including handling multiple tool calls.
  • The author compares the results favorably to Claude and says it may become their daily driver, reducing reliance on third-party hosted code providers.
  • The post is explicitly framed as anecdotal (“trust me bro”), rather than a formal benchmark or controlled study.

of course this is just a trust me bro post but I've been testing various local models (a couple gemma4s, qwen3 coder next, nemotron) and I noticed the new qwen3.6 show up on LM Studio so I hooked it up.

VERY impressed. It's super fast to respond, handles long research tasks with many tool calls (I had it investigate why R8 was breaking some serialization across an Android app), responses are on point. I think it will be my daily driver (prior was Kimi k2.5 via OpenCode zen).

FeelsGoodman, no more sending my codebase to rando providers and "trusting" them.

submitted by /u/Medical_Lengthiness6
[link] [comments]