When More Thinking Hurts: Overthinking in LLM Test-Time Compute Scaling

arXiv cs.AI / 4/14/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper challenges the common assumption that increasing LLM test-time “reasoning” (longer chains of thought) monotonically improves outcomes.
  • It shows diminishing marginal returns at higher compute budgets, including “overthinking” where extra reasoning correlates with abandoning previously correct answers.
  • The authors demonstrate that optimal thinking length depends on problem difficulty, implying that fixed/uniform compute allocation is inefficient.
  • Using a cost-aware evaluation framework, they find that stopping at moderate reasoning budgets can substantially cut computation while preserving similar accuracy.
  • Overall, the work reframes test-time compute scaling as a problem of finding an optimal stopping point rather than simply maximizing reasoning length.

Abstract

Scaling test-time compute through extended chains of thought has become a dominant paradigm for improving large language model reasoning. However, existing research implicitly assumes that longer thinking always yields better results. This assumption remains largely unexamined. We systematically investigate how the marginal utility of additional reasoning tokens changes as compute budgets increase. We find that marginal returns diminish substantially at higher budgets and that models exhibit ``overthinking'', where extended reasoning is associated with abandoning previously correct answers. Furthermore, we show that optimal thinking length varies across problem difficulty, suggesting that uniform compute allocation is suboptimal. Our cost-aware evaluation framework reveals that stopping at moderate budgets can reduce computation significantly while maintaining comparable accuracy.