StageMem: Lifecycle-Managed Memory for Language Models

arXiv cs.CL / 4/21/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that deployed LLM memory systems are often misframed as static stores, when the real challenge is dynamic memory control over time.
  • It proposes StageMem, which treats memory as a stateful process with three stages—transient, working, and durable memory—to manage retention, promotion, updating, and eviction.
  • Each memory item is modeled with explicit confidence and strength, separating low-cost initial admission from later long-term commitment.
  • The approach aims to reduce issues like retaining too many uncertain items or forgetting important content in the wrong order, improving users’ trust in what persists.
  • Experiments with “controlled pressure regimes” and adaptations of external tasks suggest the same schema can work with more robust retrieval structures beyond purely synthetic settings.

Abstract

Long-horizon language model systems increasingly rely on persistent memory, yet many current designs still treat memory primarily as a static store: write an item, place it into memory, and retrieve it later if needed. We argue that this framing does not adequately capture the practical memory-control problem in deployed LLM systems. In realistic settings, the difficulty is often not merely forgetting useful information, but retaining too many uncertain items, forgetting important content in the wrong order, and giving users little trust in what will persist over time. We propose StageMem, a lifecycle-managed memory framework that treats memory as a stateful process rather than a passive repository. StageMem organizes memory into three stages -- transient, working, and durable memory -- and models each item with explicit confidence and strength. This separates shallow admission from long-term commitment: information may first be written at low cost and only later be promoted, retained, updated, or evicted as evidence and pressure evolve. Under controlled pressure regimes, this decomposition helps preserve late-important content while keeping memory burden and deeper-tier pollution more controlled. Adapted external tasks provide boundary evidence that the same schema remains compatible with stronger retrieval structure outside pure synthetic control. We present StageMem as a principled decomposition of the memory-control problem for language model systems.