What it feels like to have to have Qwen 3.6 or Gemma 4 running locally

Reddit r/LocalLLaMA / 4/29/2026

💬 OpinionSignals & Early TrendsTools & Practical UsageModels & Research

Key Points

  • A Reddit user describes Qwen 3.6 and Gemma 4 as reliable “workhorse” local LLMs that can handle real work scenarios.
  • They say the key to success is building a system around the models’ weaknesses rather than treating the models as fully self-sufficient experts.
  • The user reports using these models to perform parts of tasks that previously required an expert billed at $200/hour, implying meaningful productivity gains.
  • They highlight strong local deployment feasibility, noting that Qwen 3.6 27B can reportedly run on a single RTX 3090.
  • The post also references earlier LLM systems (e.g., Nous Hermes 2 and Mistral) to suggest this “LLM as an expert workflow” idea has been developing for years.
What it feels like to have to have Qwen 3.6 or Gemma 4 running locally

Well or pretty close to it, they are excellent work horses. I run them in real work scenarios doing some of the work I used to do myself as an skilled expert in my field, billing 200$ an hour. Ofc the key is building a system around their weaknesses, and I've had already LLM systems doing expert work years ago when first ones came (shout out nous hermes 2 mistral!).

But yeah pretty neat, especially noonghunnas club 3090 and you can have 3.6 27B fly on a single 3090.

submitted by /u/GodComplecs
[link] [comments]