HingeMem: Boundary Guided Long-Term Memory with Query Adaptive Retrieval for Scalable Dialogues

arXiv cs.CL / 4/9/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • この論文は、対話システム向けの長期メモリ「HingeMem」を提案し、イベント区間の理論を用いて解釈可能なインデックスを構築します。
  • HingeMemは person/time/location/topic の4要素の変化(境界)をトリガーにして現在のセグメントを書き込み、冗長な更新を減らしつつ重要文脈を保持します。
  • 検索面では、問い合わせに応じて「何を」「どれだけ」取り出すかを同時に決めるクエリ適応型リトリーバルを導入し、クエリの種類推定で検索深さも制御します。
  • LLMスケール(0.6B〜本番級、例:Qwen3-0.6B〜Qwen-Flash)での実験(LOCOMO)により、強力なベースライン比で約20%相対改善を達成し、さらに HippoRAG2 比でQAトークンコストを68%削減したと報告しています。
  • 適応的検索により、長期対話で効率と信頼性を両立するWebアプリへの適用に向く手法であると位置づけています。

Abstract

Long-term memory is critical for dialogue systems that support continuous, sustainable, and personalized interactions. However, existing methods rely on continuous summarization or OpenIE-based graph construction paired with fixed Top-\textit{k} retrieval, leading to limited adaptability across query categories and high computational overhead. In this paper, we propose HingeMem, a boundary-guided long-term memory that operationalizes event segmentation theory to build an interpretable indexing interface via boundary-triggered hyperedges over four elements: person, time, location, and topic. When any such element changes, HingeMem draws a boundary and writes the current segment, thereby reducing redundant operations and preserving salient context. To enable robust and efficient retrieval under diverse information needs, HingeMem introduces query-adaptive retrieval mechanisms that jointly decide (a) \textit{what to retrieve}: determine the query-conditioned routing over the element-indexed memory; (b) \textit{how much to retrieve}: control the retrieval depth based on the estimated query type. Extensive experiments across LLM scales (from 0.6B to production-tier models; \textit{e.g.}, Qwen3-0.6B to Qwen-Flash) on LOCOMO show that HingeMem achieves approximately 20\% relative improvement over strong baselines without query categories specification, while reducing computational cost (68\%\downarrow question answering token cost compared to HippoRAG2). Beyond advancing memory modeling, HingeMem's adaptive retrieval makes it a strong fit for web applications requiring efficient and trustworthy memory over extended interactions.