What's your favorite small-medium local model?

Reddit r/LocalLLaMA / 4/16/2026

💬 OpinionSignals & Early TrendsTools & Practical Usage

Key Points

  • A Reddit user asks the community which small-to-medium local LLMs they prefer and why.
  • The post shares personal experimentation with Gemma-4-E4B and Qwen3.5-9B, including variant and quantized builds such as Gemopus, Qwopus, and an “uncensored/aggressive” Q8_0 model.
  • The user also notes that they don’t have much experience with other models, inviting recommendations from others who do.
  • The thread is positioned as discussion-driven rather than a formal benchmark or official release, focusing on subjective fit and practical local usage.

I'm now having fun with Gemma-4-E4B and Qwen3.5-9B, trying different variants like Gemopus and Qwopus, and Qwen3.5-9B-Uncensored-HauhauCS-Aggressive-Q8_0

don't quite know other models, so what's your favorite? why and how are them?

submitted by /u/__ahdw
[link] [comments]