When Continual Learning Moves to Memory: A Study of Experience Reuse in LLM Agents

arXiv cs.LG / 5/1/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Memory-augmented LLM agents can appear to enable continual learning without parameter updates, but the stability–plasticity problem still reappears at the external memory/retrieval layer.
  • With a limited context window, old and new experiences can compete during retrieval, effectively moving the continual-learning bottleneck from model updates to memory access.
  • The study proposes a (k,v) framework to separately analyze how experiences are represented and how they are organized for retrieval in external memory.
  • Experiments in ALFWorld and BabyAI show that abstract procedural memories transfer more reliably than detailed trajectories, and negative transfer tends to disproportionately affect the hardest cases.
  • Memory organization choices are not universally good: approaches that improve forward transfer can also cause severe forgetting, highlighting trade-offs in memory representation and retrieval design.

Abstract

Memory-augmented LLM agents offer an appealing shortcut to continual learning: rather than updating model parameters, they accumulate experience in external memory, seemingly sidestepping the stability-plasticity dilemma of parametric learning. We show that this challenge does not disappear but resurfaces at the memory level. Under a limited context window, old and new experiences compete during retrieval, relocating the continual-learning bottleneck from parameter updates to memory access. To study this phenomenon, we introduce a (k,v) framework that disentangles two fundamental design axes of external memory: how experience is represented and how it is organized for retrieval. Across sequential-task experiments in ALFWorld and BabyAI, we find that abstract procedural memories transfer more reliably than detailed trajectories, while negative transfer disproportionately harms the hard cases. Moreover, finer-grained memory organization is not universally beneficial: designs that yield strong forward transfer can simultaneously induce severe forgetting. Together, these results reveal that external memory does not resolve the continual-learning problem; it reshapes it into a problem of memory representation and retrieval design.