Reinforced Agent: Inference-Time Feedback for Tool-Calling Agents

arXiv cs.AI / 5/1/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper proposes “Reinforced Agent,” which brings tool-calling evaluation into the inference-time execution loop by using a specialized reviewer agent to judge provisional tool calls before they run.
  • It separates responsibilities between a primary execution agent and a secondary review agent, aiming to replace purely post-hoc error checking with proactive error mitigation.
  • The authors introduce Helpfulness-Harmfulness metrics to quantify the tradeoff between feedback that corrects base-agent mistakes (helpfulness) and feedback that degrades correct actions (harmfulness).
  • Experiments on BFCL and Tau2-Bench show measurable gains (+5.5% on irrelevance detection and +7.1% on multi-turn tasks), with reviewer model choice strongly affecting the benefit-to-risk ratio.
  • The study finds that using o3-mini as the reviewer reasoning model outperforms GPT-4o on the net ratio, and that automated prompt optimization via GEPA further improves results by about +1.5–2.8% without retraining the base agent.

Abstract

Tool-calling agents are evaluated on tool selection, parameter accuracy, and scope recognition, yet LLM trajectory assessments remain inherently post-hoc. Disconnected from the active execution loop, such assessments identify errors that are usually addressed through prompt-tuning or retraining, and fundamentally cannot course-correct the agent in real time. To close this gap, we move evaluation into the execution loop at inference time: a specialized reviewer agent evaluates provisional tool calls prior to execution, shifting the paradigm from post-hoc recovery to proactive evaluation and error mitigation. In practice, this architecture establishes a clear separation of concerns between the primary execution agent and a secondary review agent. As with any multi-agent system, the reviewer can introduce new errors while correcting others, yet no prior work to our knowledge has systematically measured this tradeoff. To quantify this tradeoff, we introduce Helpfulness-Harmfulness metrics: helpfulness measures the percentage of base agent errors that feedback corrects; harmfulness measures the percentage of correct responses that feedback degrades. These metrics directly inform reviewer design by revealing whether a given model or prompt provides net positive value. We evaluate our approach on BFCL (single-turn) and Tau2-Bench (multi-turn stateful scenarios), achieving +5.5% on irrelevance detection and +7.1% on multi-turn tasks. Our metrics reveal that reviewer model choice is critical: the reasoning model o3-mini achieves a 3:1 benefit-to-risk ratio versus 2.1:1 for GPT-4o. Automated prompt optimization via GEPA provides an additional +1.5-2.8%. Together, these results demonstrate a core advantage of separating execution and review: the reviewer can be systematically improved through model selection and prompt optimization, without retraining the base agent.