| Well or pretty close to it, they are excellent work horses. I run them in real work scenarios doing some of the work I used to do myself as an skilled expert in my field, billing 200$ an hour. Ofc the key is building a system around their weaknesses, and I've had already LLM systems doing expert work years ago when first ones came (shout out nous hermes 2 mistral!). But yeah pretty neat, especially noonghunnas club 3090 and you can have 3.6 27B fly on a single 3090. [link] [comments] |
What it feels like to have to have Qwen 3.6 or Gemma 4 running locally
Reddit r/LocalLLaMA / 4/29/2026
💬 OpinionSignals & Early TrendsTools & Practical UsageModels & Research
Key Points
- A Reddit user describes Qwen 3.6 and Gemma 4 as reliable “workhorse” local LLMs that can handle real work scenarios.
- They say the key to success is building a system around the models’ weaknesses rather than treating the models as fully self-sufficient experts.
- The user reports using these models to perform parts of tasks that previously required an expert billed at $200/hour, implying meaningful productivity gains.
- They highlight strong local deployment feasibility, noting that Qwen 3.6 27B can reportedly run on a single RTX 3090.
- The post also references earlier LLM systems (e.g., Nous Hermes 2 and Mistral) to suggest this “LLM as an expert workflow” idea has been developing for years.
Related Articles

Black Hat USA
AI Business
LLMs will be a commodity
Reddit r/artificial
Indian Developers: How to Build AI Side Income with $0 Capital in 2026
Dev.to
HubSpot Just Legitimized AEO: What It Means for Your Brand AI Visibility
Dev.to

From Fault Codes to Smart Fixes: How Google Cloud NEXT ’26 Inspired My AI Mechanic Assistant
Dev.to