AutoSearch: Adaptive Search Depth for Efficient Agentic RAG via Reinforcement Learning

arXiv cs.AI / 4/21/2026

📰 NewsIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • Agentic RAG systems can waste computation and increase latency due to redundant multi-step retrieval searches, so prior approaches often cap search depth to reduce cost.
  • The paper analyzes the relationship between search depth, question complexity, and the agent’s capability, and identifies a “minimal sufficient” depth that balances accuracy and efficiency.
  • AutoSearch is introduced as a reinforcement learning framework that judges each retrieval step by generating self-produced intermediate answers.
  • By rewarding the achievement of the minimal sufficient depth and penalizing over-searching (with additional stabilizing reward design), AutoSearch reduces unnecessary steps while maintaining answer quality.
  • Experiments across multiple benchmarks show AutoSearch improves the accuracy–efficiency trade-off by alleviating over-searching.

Abstract

Agentic retrieval-augmented generation (RAG) systems enable large language models (LLMs) to solve complex tasks through multi-step interaction with external retrieval tools. However, such multi-step interaction often involves redundant search steps, incurring substantial computational cost and latency. Prior work limits search depth (i.e., the number of search steps) to reduce cost, but this often leads to underexploration of complex questions. To address this, we first investigate how search depth affects accuracy and find a minimal sufficient search depth that defines an accuracy-efficiency trade-off, jointly determined by question complexity and the agent's capability. Furthermore, we propose AutoSearch, a reinforcement learning (RL) framework that evaluates each search step via self-generated intermediate answers. By a self-answering mechanism, AutoSearch identifies the minimal sufficient search depth and promotes efficient search by rewarding its attainment while penalizing over-searching. In addition, reward mechanisms are introduced to stabilize search behavior and improve answer quality on complex questions. Extensive experiments on multiple benchmarks show that AutoSearch achieves a superior accuracy-efficiency trade-off, alleviating over-searching while preserving search quality.