SCM: Sleep-Consolidated Memory with Algorithmic Forgetting for Large Language Models

arXiv cs.LG / 4/24/2026

📰 NewsModels & Research

Key Points

  • The paper proposes SCM (Sleep-Consolidated Memory), a new memory architecture for large language models aimed at providing persistent, structured, and more biologically plausible memory than current context-window and storage approaches.
  • SCM combines limited-capacity working memory, multi-dimensional importance tagging, offline “sleep”-stage consolidation (with distinct NREM and REM phases), value-based intentional forgetting, and a computational self-model for introspection.
  • On an 8-test benchmark suite, the prototype reportedly reaches perfect recall accuracy on ten-turn conversations while cutting memory noise by 90.9% via adaptive forgetting.
  • The approach maintains memory search latency under 1 millisecond even when storing hundreds of concepts, and the authors position it as a testable foundation for future LLM memory research.

Abstract

We present SCM (Sleep-Consolidated Memory), a research preview of a memory architecture for large language models that draws on neuroscientific principles to address a fundamental limitation in current systems: the absence of persistent, structured, and biologically plausible memory. Existing approaches rely on truncating context windows, growing vector databases without bound, or tiered storage systems that lack consolidation and forgetting mechanisms. SCM implements five core components inspired by human memory: a limited-capacity working memory, multi-dimensional importance tagging, offline sleep-stage consolidation with distinct NREM and REM phases, intentional value-based forgetting, and a computational self-model enabling introspection. Across a standardized benchmark suite of eight tests, the prototype achieves perfect recall accuracy over ten-turn conversations while reducing memory noise by 90.9% through adaptive forgetting. Memory search latency remains below one millisecond even with hundreds of stored concepts. This work establishes the architectural foundations for memory systems that consolidate, prioritize, and forget, offering a testable platform for advancing LLM memory research.