How to increase coding ability in smaller models?

Reddit r/LocalLLaMA / 4/19/2026

💬 OpinionDeveloper Stack & InfrastructureTools & Practical Usage

Key Points

  • The author is using a smaller Qwen3.5 35B model via opencode to generate software code, but the output often introduces errors that require significant debugging time.
  • They ask whether there are specific plugins, protocols, or workflow adjustments that can improve the model’s coding quality and reduce the rate of broken implementations.
  • They are open to switching to a different model if it offers a better quality-to-speed tradeoff than their current setup (about 30 tokens/second).
  • The post includes their local hardware specs (RTX 4070 12GB, Ryzen 7 5800X3D, 32GB DDR4 RAM) to contextualize the model choice and performance constraints.

I've been running Qwen3.5 35b APEX I Quality to code a piece of software for me through opencode. Are there any plugins/protocols I should be using to give it better coding skills? It constantly messing things up so 90% of the time spent is tracking down issues its created. Also open to using a different model. I've just found this has been the best quality/speed ratio. Currently getting around 30t/s.

System specs:

RTX 4070 12GB

RYZEN 7 5800X3D

32GB DDR4 RAM

submitted by /u/keepthememes
[link] [comments]