Memory Intelligence Agent

arXiv cs.AI / 4/7/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The article introduces a new “Memory Intelligence Agent” (MIA) framework for deep research agents that use LLM reasoning plus external tools, aiming to make memory use more efficient and evolution-capable.
  • MIA uses a Manager-Planner-Executor architecture where a non-parametric memory manager stores compressed historical trajectories, a planner (parametric memory agent) generates search plans, and an executor follows those plans to search and analyze.
  • It addresses limitations of prior memory-trajectory retrieval approaches by adding an alternating reinforcement learning setup for better Planner–Executor cooperation and by enabling test-time learning where the planner evolves on-the-fly during inference.
  • The framework further improves memory evolution via a bidirectional conversion loop between parametric and non-parametric memories, plus reflection and unsupervised judgment to support reasoning and self-evolution in open-ended settings.
  • Experiments across eleven benchmarks reportedly show that MIA outperforms existing methods, highlighting benefits for autonomous, tool-using LLM agents that rely on long-term experience.

Abstract

Deep research agents (DRAs) integrate LLM reasoning with external tools. Memory systems enable DRAs to leverage historical experiences, which are essential for efficient reasoning and autonomous evolution. Existing methods rely on retrieving similar trajectories from memory to aid reasoning, while suffering from key limitations of ineffective memory evolution and increasing storage and retrieval costs. To address these problems, we propose a novel Memory Intelligence Agent (MIA) framework, consisting of a Manager-Planner-Executor architecture. Memory Manager is a non-parametric memory system that can store compressed historical search trajectories. Planner is a parametric memory agent that can produce search plans for questions. Executor is another agent that can search and analyze information guided by the search plan. To build the MIA framework, we first adopt an alternating reinforcement learning paradigm to enhance cooperation between the Planner and the Executor. Furthermore, we enable the Planner to continuously evolve during test-time learning, with updates performed on-the-fly alongside inference without interrupting the reasoning process. Additionally, we establish a bidirectional conversion loop between parametric and non-parametric memories to achieve efficient memory evolution. Finally, we incorporate a reflection and an unsupervised judgment mechanisms to boost reasoning and self-evolution in the open world. Extensive experiments across eleven benchmarks demonstrate the superiority of MIA.