AI Navigate

NextMem: Towards Latent Factual Memory for LLM-based Agents

arXiv cs.AI / 3/18/2026

📰 NewsModels & Research

Key Points

  • NextMem presents a latent factual memory framework for LLM-based agents to improve memory efficiency and retrieval over textual or parametric approaches.
  • It employs an autoregressive autoencoder to construct latent memory with accurate reconstruction of past observations.
  • The training pipeline includes autoregressive reconstruction alignment and progressive latent substitution, along with quantization to reduce storage overhead.
  • Experiments show improved retrieval, robustness, and extensibility, and the authors release code and checkpoints on GitHub.

Abstract

Memory is critical for LLM-based agents to preserve past observations for future decision-making, where factual memory serves as its foundational part. However, existing approaches to constructing factual memory face several limitations. Textual methods impose heavy context and indexing burdens, while parametric methods suffer from catastrophic forgetting and high costs. To address these challenges, we introduce NextMem, a latent factual memory framework that utilizes an autoregressive autoencoder to efficiently construct latent memory while ensuring accurate reconstruction. For better optimization, we propose a two-stage training process, including autoregressive reconstruction alignment and progressive latent substitution. We also incorporate quantization to reduce storage overhead. Extensive experiments demonstrate that NextMem achieves superior performance, and excels in retrieval, robustness, and extensibility properties. We release our code and model checkpoints at https://github.com/nuster1128/NextMem.