ART: Attention Replacement Technique to Improve Factuality in LLMs

arXiv cs.CL / 4/9/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper investigates how attention patterns inside LLM layers and heads relate to hallucinations, finding that shallow layers often use uniform, evenly distributed attention that can cause the model to miss relevant context.
  • It introduces a training-free method called Attention Replacement Technique (ART) that swaps shallow-layer uniform attention with local attention to encourage focus on pertinent segments.
  • Experiments across multiple LLM architectures show ART significantly reduces hallucinations in question answering and related factuality-sensitive tasks.
  • The approach is designed to be generalizable and effective without fine-tuning or requiring additional training data, making it easier to adopt in existing deployments.
  • Overall, the work reframes hallucination mitigation as something influenced by internal attention behavior rather than only output-level decoding tricks.

Abstract

Hallucination in large language models (LLMs) continues to be a significant issue, particularly in tasks like question answering, where models often generate plausible yet incorrect or irrelevant information. Although various methods have been proposed to mitigate hallucinations, the relationship between attention patterns and hallucinations has not been fully explored. In this paper, we analyze the distribution of attention scores across each layer and attention head of LLMs, revealing a common and intriguing phenomenon: shallow layers of LLMs primarily rely on uniform attention patterns, where the model distributes its attention evenly across the entire sequence. This uniform attention pattern can lead to hallucinations, as the model fails to focus on the most relevant information. To mitigate this issue, we propose a training-free method called Attention Replacement Technique (ART), which replaces these uniform attention patterns in the shallow layers with local attention patterns. This change directs the model to focus more on the relevant contexts, thus reducing hallucinations. Through extensive experiments, ART demonstrates significant reductions in hallucinations across multiple LLM architectures, proving its effectiveness and generalizability without requiring fine-tuning or additional training data.