AMD's AI director slams Claude Code for becoming dumber and lazier since last update
'Claude cannot be trusted to perform complex engineering tasks' according to GitHub ticket
If you've noticed Claude Code's performance degrading to the point where you find you don't trust it to handle complicated tasks anymore, you're not alone.
A GitHub issue was filed on Friday by user stellaraccident. That user's Github profile and a related LinkedIn post identify the poster as Stella Laurenzo, the director of the AI group at chipmaker AMD. She complains that, ever since some time in February, Claude Code has really been phoning it in.
"Claude cannot be trusted to perform complex engineering tasks," Laurenzo wrote, noting that her team reached that conclusion by referring to months of logs from the "very consistent, high complexity work environment" in which they use Claude Code. "Every senior engineer on my team has reported similar experiences/anecdotes," Laurenzo added.
Based on comments in the issue thread, plenty of others are feeling the same way, and Reddit commenters have expressed similar sentiments.
To reach this conclusion, Laurenzo and her team analyzed 6,852 Claude Code sessions incorporating 234,760 tool calls and 17,871 thinking blocks. According to their data, the number of stop-hook violations used to catch ownership dodging, premature cessation of the thinking process, and permission-seeking behavior that indicate "laziness" all skyrocketed, going from zero prior to the 8th of March to 10 per day on average through the end of last month.
The number of times Claude would read through a piece of code before making changes also dropped drastically, going from 6.6 reads on average to just 2 by the end of March, while over the same period, Claude began rewriting entire files instead of making edits with much greater frequency.
All of those things, said Laurenzo, point to Claude Code not thinking as deeply, and coincide with the early March deployment of thinking content redaction with Claude Code version 2.1.69. Thinking redaction functions as a header that defaults to stripping thinking content from Claude Code API responses, meaning users don't get any idea what Claude Code is actually doing while it reflects on a request.
The evidence, according to Lorenzo, points to a general thinking reduction since the implementation.
"When thinking is shallow, the model defaults to the cheapest action available: edit without reading, stop without finishing, dodge responsibility for failures, take the simplest fix rather than the correct one," the GitHub issue explains. "These are exactly the symptoms observed."
If you're wondering, this appears to be a separate issue from the issue Claude Code users cried foul about back in February when version 2.1.20 of the bot caused it to truncate its explanation of what it was reading as part of its thinking process.
In that instance, which led many Claude Code users to decry that it was evidence the AI was being dumbed down, users were left with just a brief line indicating how many files were read with little more specificity than that. We can't imagine those same developers will be very happy about this latest development.
Anthropic has also caught flak for unexplained surges in token usage that have pushed some users past their limits, leaving them unable to use the product. Add to that the recent exposure of Claude Code's entire source code, and it's not looking good for the AI firm.
For Laurenzo's part, she wants Anthropic to be transparent about whether it's reducing or capping thinking tokens and causing Claude Code to vomit garbage. At the very least, she wants Claude to expose the number of thinking tokens being used per request to let users "monitor whether their requests are getting the reasoning depth they need."
Laurenzo also asked for a max thinking tier to be added to Anthropic's offerings for engineers running complex workflows. "The current subscription model doesn't distinguish between users who need 200 thinking tokens per response and users who need 20,000," the AMD AI chief explained. "Users running complex engineering workflows would pay significantly more for guaranteed deep thinking."
- Claude Code source leak reveals how much info Anthropic can hoover up about you and your system
- Using AI to code does not mean your code is more secure
- Anthropic sure has a mess on its hands thanks to that Claude Code source leak
- Claude Code's prying AIs read off-limits secret files
"We have switched to another provider which is doing superior quality work, but Claude has been good to us, and we are leaving this in the hopes that Anthropic can fix their product," Laurenzo explained, while declining to go into details in a comment citing NDAs about whatever new tool her team is using. That said, Laurenzo did warn Anthropic that it's still early in the AI coding game and Anthropic is looking at giving up the top spot if its behavior continues.
"All I will add is that 6 months ago, Claude stood alone in terms of reasoning quality and execution," Laurenzo added in a response on the issue thread. "But the others need to be watched and evaluated very carefully. Anthropic is far from alone at the capability tier that Opus previously occupied."
Neither Anthropic nor Laurenzo initially responded to questions for this story. ®




