Ask Only When Needed: Proactive Retrieval from Memory and Skills for Experience-Driven Lifelong Agents

arXiv cs.CL / 4/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that most lifelong agents use retrieval passively, so they often miss knowledge gaps during an ongoing task and retrieve too late or too much.
  • It proposes ProactAgent, which supports proactive retrieval over a structured experience base that separates factual memory, episodic memory, and behavioral skills.
  • ProactAgent includes ExpOnEvo for continual improvement via both policy updates and memory refinement, enabling ongoing evolution of both behavior and stored experience.
  • It introduces ProactRL, treating retrieval as an explicit policy action and training it with step-level supervision by comparing outcomes with and without retrieval from the same interaction prefix.
  • Experiments on SciWorld, AlfWorld, and StuLife show higher success rates (73.50% on SciWorld and 71.28% on AlfWorld) alongside lower retrieval overhead, with competitive performance on StuLife.

Abstract

Online lifelong learning enables agents to accumulate experience across interactions and continually improve on long-horizon tasks. However, existing methods typically treat retrieval from past experience as a passive operation, triggering it only at task initialization or after completing a step. Consequently, agents often fail to identify knowledge gaps during interaction and proactively retrieve the most useful experience for the current decision. To address this limitation, we present ProactAgent, an experience-driven lifelong learning framework for proactive retrieval over a structured experience base. We first introduce Experience-Enhanced Online Evolution (ExpOnEvo), which enables continual improvement through both policy updates and memory refinement. The experience base organizes historical interactions into typed repositories, including factual memory, episodic memory, and behavioral skills, so that retrieval can provide both relevant evidence and actionable guidance. On top of this, we propose Proactive Reinforcement Learning-based Retrieval (ProactRL), which models retrieval as an explicit policy action and learns when and what to retrieve via paired-branch process rewards. By comparing continuations from identical interaction prefixes with and without retrieval, ProactRL provides step-level supervision for retrieval decisions, encouraging retrieval only when it leads to better task outcomes or higher efficiency. Experiments on SciWorld, AlfWorld, and StuLife show that ProactAgent consistently improves lifelong agent performance, achieving success rates of 73.50\% on SciWorld and 71.28\% on AlfWorld while substantially reducing retrieval overhead, and attains performance competitive with proprietary models on StuLife.