COSEARCH: Joint Training of Reasoning and Document Ranking via Reinforcement Learning for Agentic Search

arXiv cs.AI / 4/21/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Agentic search has improved using reinforcement learning, but prior work often leaves the document retrieval/ranking component fixed while optimizing only the reasoning agent.
  • The paper reports that replacing a fixed retrieval system with an oracle can yield up to a +26.8% relative F1 gain across seven QA benchmarks, indicating retrieval is a major bottleneck.
  • It proposes CoSearch, which jointly trains a multi-step reasoning agent and a generative document ranker using Group Relative Policy Optimization (GRPO).
  • To make GRPO work for the ranker despite variable inputs across reasoning trajectories, the authors introduce a semantic grouping method that clusters sub-queries by token-level similarity without extra rollouts.
  • Experiments on seven single-hop and multi-hop QA benchmarks show consistent improvements over strong baselines, and ablations confirm the contribution of each component, supporting joint training as a key ingredient for future search agents.

Abstract

Agentic search -- the task of training agents that iteratively reason, issue queries, and synthesize retrieved information to answer complex questions -- has achieved remarkable progress through reinforcement learning (RL). However, existing approaches such as Search-R1, treat the retrieval system as a fixed tool, optimizing only the reasoning agent while the retrieval component remains unchanged. A preliminary experiment reveals that the gap between an oracle and a fixed retrieval system reaches up to +26.8% relative F1 improvement across seven QA benchmarks, suggesting that the retrieval system is a key bottleneck in scaling agentic search performance. Motivated by this finding, we propose CoSearch, a framework that jointly trains a multi-step reasoning agent and a generative document ranking model via Group Relative Policy Optimization (GRPO). To enable effective GRPO training for the ranker -- whose inputs vary across reasoning trajectories -- we introduce a semantic grouping strategy that clusters sub-queries by token-level similarity, forming valid optimization groups without additional rollouts. We further design a composite reward combining ranking quality signals with trajectory-level outcome feedback, providing the ranker with both immediate and long-term learning signals. Experiments on seven single-hop and multi-hop QA benchmarks demonstrate consistent improvements over strong baselines, with ablation studies validating each design choice. Our results show that joint training of the reasoning agent and retrieval system is both feasible and strongly performant, pointing to a key ingredient for future search agents.