HyperMem: Hypergraph Memory for Long-Term Conversations

arXiv cs.CL / 4/10/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • HyperMemは、長期会話に必要な記憶を、従来のRAGやグラフ記憶の「要素間の単純な関係」ではなく「複数要素の同時依存(高次関係)」まで明示的に扱うハイパーグラフ型メモリとして提案しています。
  • 記憶は3階層(topics / episodes / facts)で整理され、関連するエピソードと事実をハイパーエッジで束ねて散在情報を一貫した単位として結び直します。
  • さらに、ハイブリッドな語彙・セマンティックのインデックスと、粗い→細かい(coarse-to-fine)検索戦略により、高次アソシエーションを高精度かつ効率的に取得できる設計になっています。
  • LoCoMoベンチマークの実験では、LLM-as-a-judge精度92.73%を達成し、既存手法を上回る(state-of-the-art)結果が報告されています。

Abstract

Long-term memory is essential for conversational agents to maintain coherence, track persistent tasks, and provide personalized interactions across extended dialogues. However, existing approaches as Retrieval-Augmented Generation (RAG) and graph-based memory mostly rely on pairwise relations, which can hardly capture high-order associations, i.e., joint dependencies among multiple elements, causing fragmented retrieval. To this end, we propose HyperMem, a hypergraph-based hierarchical memory architecture that explicitly models such associations using hyperedges. Particularly, HyperMem structures memory into three levels: topics, episodes, and facts, and groups related episodes and their facts via hyperedges, unifying scattered content into coherent units. Leveraging this structure, we design a hybrid lexical-semantic index and a coarse-to-fine retrieval strategy, supporting accurate and efficient retrieval of high-order associations. Experiments on the LoCoMo benchmark show that HyperMem achieves state-of-the-art performance with 92.73% LLM-as-a-judge accuracy, demonstrating the effectiveness of HyperMem for long-term conversations.