Information Gain-based Policy Optimization: A Simple and Effective Approach for Multi-Turn Search Agents

arXiv cs.CL / 3/25/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that RL training for LLM search agents suffers from reward sparsity in multi-turn settings when supervision is only given after the final answer is produced.
  • It introduces Information Gain-based Policy Optimization (IGPO), which gives dense, turn-level rewards by estimating marginal increases in the model’s probability of producing the correct answer as each interaction turn progresses.
  • IGPO derives intrinsic supervision directly from the model’s own belief updates, avoiding reliance on external reward models or expensive Monte Carlo estimation used by some prior approaches.
  • Experiments on in-domain and out-of-domain multi-turn search benchmarks show IGPO improves accuracy and sample efficiency compared with strong baselines.
  • The authors provide an open-source implementation to support reproduction and adoption of the method for multi-turn agent training.

Abstract

Large language model (LLM)-based agents are increasingly trained with reinforcement learning (RL) to enhance their ability to interact with external environments through tool use, particularly in search-based settings that require multi-turn reasoning and knowledge acquisition. However, existing approaches typically rely on outcome-based rewards that are only provided exclusively upon generating the final answer. This reward sparsity becomes particularly problematic in multi-turn settings, where long trajectories exacerbate three critical issues: (i) advantage collapse, where all rollouts receive identical rewards and provide no useful learning signals; (ii) lack of fine-grained credit assignment, where the correctness of intermediate turns is obscured, especially in long-horizon tasks; and (iii) poor sample efficiency, where each rollout yields only a single outcome signal, leading to low data utilization. In this paper, we propose Information Gain-based Policy Optimization (IGPO), a simple yet effective RL framework that provides dense and intrinsic supervision for multi-turn agent training. IGPO models each interaction turn as an incremental process of acquiring information about the ground truth, and defines turn-level rewards as the marginal increase in the policy's probability of producing the correct answer. Unlike prior process-level reward approaches that depend on external reward models or costly Monte Carlo estimation, IGPO derives intrinsic rewards directly from the model's own belief updates. These intrinsic turn-level rewards are combined with outcome-level supervision to form dense reward signals. Extensive experiments on both in-domain and out-of-domain benchmarks demonstrate that IGPO consistently outperforms strong baselines in multi-turn scenarios, achieving higher accuracy and improved data efficiency. Our code is available at https://github.com/GuoqingWang1/IGPO.