| TL;DR:
In each of these they made conscious choices to lower server load at the cost of quality, completely outside the end users control and without informing their paying customers of the changes. For me, this proves that if you depend on an AI model for your service or to do your job, the only sane choice is to pick an open-weight model that you can host yourself, or that you can pay someone to host for you. [link] [comments] |
Anthropic admits to have made hosted models more stupid, proving the importance of open weight, local models
Reddit r/LocalLLaMA / 4/24/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisIndustry & Market Moves
Key Points
- Anthropic says it adjusted Claude Code’s default reasoning effort from “high” to “medium” to reduce long latency and UI freezes, but reverted the change after users preferred higher intelligence with lower effort for simple tasks.
- A later update intended to clear old “thinking” from idle sessions reduced resume latency, but a bug caused the behavior to repeat every turn and made Claude seem repetitive; Anthropic fixed it.
- Anthropic also added a system prompt to reduce verbosity, but the combination with other prompt changes degraded coding quality, leading to a rollback.
- The changes affected multiple Claude 4.x variants (Sonnet 4.6 and Opus 4.6, plus Opus 4.7) and are framed as tradeoffs that reduced quality in ways end users may not control.
- The takeaway highlighted by commenters is that service reliability and control may favor using open-weight models that users can host (or pay others to host) rather than relying solely on opaque hosted behavior changes.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.




