Is there anything better than Qwen3.5-27B-UD-Q5_K_XL for coding?

Reddit r/LocalLLaMA / 4/14/2026

💬 OpinionSignals & Early TrendsTools & Practical Usage

Key Points

  • A Reddit user with a 32GB VRAM limit (RTX 5090) reports that Qwen3.5-27B-UD-Q5_K_XL, using opencode and mmproj, performs well for their mainly web-development coding tasks.
  • They note they sometimes use Claude and Codex, but rely on Qwen more recently due to tightened usage limits, switching only when Qwen gets stuck or starts repeating.
  • The post asks the community whether there are better local models to try within similar hardware constraints, or upcoming models worth tracking.
  • Overall, the discussion centers on practical model selection and workflow tradeoffs for local coding assistants rather than any new release.

I have a 5090, so my VRAM is limited to 32GB, but i find that Qwen3.5-27B-UD-Q5_K_XL with opencode (and mmproj) does a pretty good job for my use case (mainly web development).

i use claude and codex here and there, recently a lot less, because usage limits got nerfed hard. really only when qwen gets stuck or repeats himself over and over again, which happens, but sometimes i'm too lazy to be more specific and spin up claude or codex.

is there any other model i should try? or is there something coming out i should have on my radar?

submitted by /u/hedsht
[link] [comments]