Multi-Layered Memory Architectures for LLM Agents: An Experimental Evaluation of Long-Term Context Retention

arXiv cs.CV / 4/1/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a Multi-Layer Memory Framework for LLM agents that separates dialogue history into working, episodic, and semantic layers to address long-horizon semantic drift and unstable retention across sessions.
  • It uses adaptive retrieval gating and retention regularization to keep cross-session memory stable while bounding context growth and maintaining computational efficiency.
  • Experiments on LOCOMO, LOCCO, and LoCoMo report improved long-term dialogue and reasoning outcomes, including a 46.85 success rate and 0.618 overall F1 (with 0.594 multi-hop F1).
  • The approach improves six-period retention to 56.90%, reduces false memory rate to 5.1%, and lowers context usage to 58.40%, indicating better memory quality under constrained context budgets.

Abstract

Long-horizon dialogue systems suffer from semanticdrift and unstable memory retention across extended sessions. This paper presents a Multi-Layer Memory Framework that decomposes dialogue history into working, episodic, and semantic layers with adaptive retrieval gating and retention regularization. The architecture controls cross-session drift while maintaining bounded context growth and computational efficiency. Experiments on LOCOMO, LOCCO, and LoCoMo show improved performance, achieving 46.85 Success Rate, 0.618 overall F1 with 0.594 multi-hop F1, and 56.90% six-period retention while reducing false memory rate to 5.1% and context usage to 58.40%. Results confirm enhanced long-term retention and reasoning stability under constrained context budgets.

Multi-Layered Memory Architectures for LLM Agents: An Experimental Evaluation of Long-Term Context Retention | AI Navigate