SinkTrack: Attention Sink based Context Anchoring for Large Language Models

arXiv cs.CV / 4/14/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • SinkTrack is a proposed context-anchoring method for LLMs that leverages the intrinsic “attention sink” behavior that tends to keep high attention on the <BOS> token throughout generation.
  • The method injects key contextual features (e.g., from the input instruction or image) into the <BOS> representation to reduce attention drift, thereby mitigating hallucination and context forgetting.
  • SinkTrack is training-free, plug-and-play, and adds negligible inference overhead, making it practical to integrate into existing LLM pipelines.
  • Reported experiments show consistent improvements on both text and multimodal benchmarks (e.g., +21.6% on SQuAD2.0 with Llama3.1-8B-Instruct and +22.8% on M3CoT with Qwen2.5-VL-7B-Instruct) across architectures and scales.
  • The paper includes an analysis of the mechanism in terms of information delivery and provides open-source code.

Abstract

Large language models (LLMs) suffer from hallucination and context forgetting. Prior studies suggest that attention drift is a primary cause of these problems, where LLMs' focus shifts towards newly generated tokens and away from the initial input context. To counteract this, we make use of a related, intrinsic characteristic of LLMs: attention sink -- the tendency to consistently allocate high attention to the very first token (i.e., ) of a sequence. Concretely, we propose an advanced context anchoring method, SinkTrack, which treats as an information anchor and injects key contextual features (such as those derived from the input image or instruction) into its representation. As such, LLM remains anchored to the initial input context throughout the entire generation process. SinkTrack is training-free, plug-and-play, and introduces negligible inference overhead. Experiments demonstrate that SinkTrack mitigates hallucination and context forgetting across both textual (e.g., +21.6% on SQuAD2.0 with Llama3.1-8B-Instruct) and multi-modal (e.g., +22.8% on M3CoT with Qwen2.5-VL-7B-Instruct) tasks. Its consistent gains across different architectures and scales underscore the robustness and generalizability. We also analyze its underlying working mechanism from the perspective of information delivery. Our source code is available at https://github.com/67L1/SinkTrack.