AI Navigate

Beyond Sequential Distance: Inter-Modal Distance Invariant Position Encoding

arXiv cs.CV / 3/12/2026

📰 NewsModels & Research

Key Points

  • The paper identifies that the distance-based inductive bias of Multimodal RoPE degrades inter-modal attention as the text sequence length increases, causing visual fading in long-context generation.
  • It proposes inter-modal Distance Invariant Position Encoding (DIPE), which disentangles position encoding by modality to preserve intra-modal locality while anchoring inter-modal proximity.
  • DIPE, when combined with Multimodal RoPE, mitigates the inter-modal distance penalty and keeps visual signals perceptually grounded across long contexts.
  • Experimental results show preserved performance on short-context benchmarks alongside significantly improved long-context visual grounding, with code available at the linked GitHub repository.

Abstract

Despite the remarkable capabilities of Multimodal Large Language Models (MLLMs), they still suffer from visual fading in long-context scenarios. Specifically, the attention to visual tokens diminishes as the text sequence lengthens, leading to text generation detached from visual constraints. We attribute this degradation to the inherent inductive bias of Multimodal RoPE, which penalizes inter-modal attention as the distance between visual and text tokens increases. To address this, we propose inter-modal Distance Invariant Position Encoding (DIPE), a simple but effective mechanism that disentangles position encoding based on modality interactions. DIPE retains the natural relative positioning for intra-modal interactions to preserve local structure, while enforcing an anchored perceptual proximity for inter-modal interactions. This strategy effectively mitigates the inter-modal distance-based penalty, ensuring that visual signals remain perceptually consistent regardless of the context length. Experimental results demonstrate that by integrating DIPE with Multimodal RoPE, the model maintains stable visual grounding in long-context scenarios, significantly alleviating visual fading while preserving performance on standard short-context benchmarks. Code is available at https://github.com/lchen1019/DIPE.