StateLinFormer: Stateful Training Enhancing Long-term Memory in Navigation

arXiv cs.LG / 2026/3/26

📰 ニュースSignals & Early TrendsIdeas & Deep AnalysisModels & Research

要点

  • The paper introduces StateLinFormer, a stateful linear-attention navigation model designed to retain long-term memory across training segments rather than resetting at batch boundaries.
  • By preserving recurrent memory states, the authors propose a training method that better approximates learning on infinitely long sequences and supports long-horizon memory retention.
  • Experiments in MAZE and ProcTHOR show that StateLinFormer significantly outperforms both its stateless linear-attention variant and fixed-context-window Transformer baselines.
  • The results indicate that as interaction length grows, stateful training improves context-dependent adaptation, implying stronger in-context learning (ICL)-like capabilities for navigation.

Abstract

Effective navigation intelligence relies on long-term memory to support both immediate generalization and sustained adaptation. However, existing approaches face a dilemma: modular systems rely on explicit mapping but lack flexibility, while Transformer-based end-to-end models are constrained by fixed context windows, limiting persistent memory across extended interactions. We introduce StateLinFormer, a linear-attention navigation model trained with a stateful memory mechanism that preserves recurrent memory states across consecutive training segments instead of reinitializing them at each batch boundary. This training paradigm effectively approximates learning on infinitely long sequences, enabling the model to achieve long-horizon memory retention. Experiments across both MAZE and ProcTHOR environments demonstrate that StateLinFormer significantly outperforms its stateless linear-attention counterpart and standard Transformer baselines with fixed context windows. Notably, as interaction length increases, persistent stateful training substantially improves context-dependent adaptation, suggesting an enhancement in the model's In-Context Learning (ICL) capabilities for navigation tasks.