Single-Agent LLMs Outperform Multi-Agent Systems on Multi-Hop Reasoning Under Equal Thinking Token Budgets
arXiv cs.CL / 4/6/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues, using an information-theoretic view based on the Data Processing Inequality, that with a fixed reasoning-token budget and perfect context utilization, single-agent LLMs (SAS) should be at least as information-efficient as multi-agent systems (MAS) for multi-hop reasoning.
- It predicts MAS only becomes competitive when single-agent context utilization is degraded or when MAS is allowed to use more compute than SAS.
- In a controlled study across three model families (Qwen3, DeepSeek-R1-Distill-Llama, and Gemini 2.5), the authors find SAS consistently match or outperform MAS on multi-hop reasoning when reasoning tokens are held constant.
- The analysis identifies evaluation artifacts—especially in API-based budget control for Gemini 2.5 and in standard benchmarks—that can falsely inflate MAS advantages.
- The authors conclude that many reported gains from multi-agent approaches are largely explained by unaccounted computation and context effects rather than inherent architectural benefits, emphasizing the need to explicitly control compute–context–coordination trade-offs.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Black Hat Asia
AI Business

How Bash Command Safety Analysis Works in AI Systems
Dev.to

How I Built an AI Agent That Earns USDC While I Sleep — A Complete Guide
Dev.to

How to Get Better Output from AI Tools (Without Burning Time and Tokens)
Dev.to

How I Added LangChain4j Without Letting It Take Over My Spring Boot App
Dev.to