Qwen3.6-35B is worse at tool use and reasoning loops than 3.5?

Reddit r/LocalLLaMA / 4/17/2026

💬 OpinionSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • A user reports that Qwen3.6-35B, when tested locally (via oMLX and LM Studio) on coding and agent-based tasks, appears to loop through “reasoning” more often than Qwen3.5.
  • They also observe occasional failed tool calls, though they suspect some of those failures could stem from tool/parsing issues rather than the model itself.
  • The model reportedly performs acceptably on simpler applications but struggles with more complex, iterative work like progressing on a basic 3D game even with minimal context.
  • The user states the looping/tool issues persisted across multiple quantization formats and settings (including 8-bit MLX, Q6_K_XL, Q8_XL, and BF16), suggesting the behavior may be inherent to the model.
  • The post is a call for others to confirm whether they see similar regressions in tool use and reasoning-loop behavior.

Been running the new model entire evening in different quants and coding tasks with OpenCode. Used oMLX and LM Studio. Used recommended settings for precise tasks (temp 0.6, top-k 20, etc) and OpenCode agent. So far my findings is that the model goes into infinite reasoning loops more often than 3.5, and I sometimes see failed tool calls. The latter could be parser bugs, but the former is the model itself.

It’s ok on basic apps, but really struggles to move ahead on something more complex like a simple 3D game even when the context is nearly empty, as if it tries to be super defensive and rechecks itself continuously.

Does anyone else have similar observations?

Edit: forgot to mention I tried 8bit MLX, Q6_K_XL, Q8_XL, BF16, all had this problem

submitted by /u/mr_il
[link] [comments]