AgentCollab: A Self-Evaluation-Driven Collaboration Paradigm for Efficient LLM Agents

arXiv cs.CL / 3/30/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes AgentCollab, a self-evaluation-driven framework for coordinating LLM agents across multiple reasoning-capability tiers to balance execution efficiency and robustness during long-horizon tasks.
  • Instead of external routing, AgentCollab uses the agent’s own self-reflection signal to judge whether the current reasoning path is making meaningful progress and escalates to a stronger model only when needed.
  • It adds a difficulty-aware cumulative escalation strategy that increases allocated reasoning budget based on recent failure signals to stabilize performance over extended multi-step interactions.
  • Using a two-level small/large model setup, the experiments on multiple multi-step agent benchmarks show improved accuracy-efficiency trade-offs versus baseline approaches, strengthening the Pareto frontier.

Abstract

Autonomous agents powered by large language models (LLMs) perform complex tasks through long-horizon reasoning and tool interaction, where a fundamental trade-off arises between execution efficiency and reasoning robustness. Models at different capability-cost levels offer complementary advantages: lower-cost models enable fast execution but may struggle on difficult reasoning segments, while stronger models provide more robust reasoning at higher computational cost. We present AgentCollab, a self-driven collaborative inference framework that dynamically coordinates models with different reasoning capacities during agent execution. Instead of relying on external routing modules, the framework uses the agent's own self-reflection signal to determine whether the current reasoning trajectory is making meaningful progress, and escalates control to a stronger reasoning tier only when necessary. To further stabilize long-horizon execution, we introduce a difficulty-aware cumulative escalation strategy that allocates additional reasoning budget based on recent failure signals. In our experiments, we instantiate this framework using a two-level small-large model setting. Experiments on diverse multi-step agent benchmarks show that AgentCollab consistently improves the accuracy-efficiency Pareto frontier of LLM agents.