System-Mediated Attention Imbalances Make Vision-Language Models Say Yes

arXiv cs.CL / 4/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper links common vision-language model (VLM) “yes-bias” hallucinations to imbalanced attention allocation across system, image, and text modalities.
  • It argues that prior fixes often treat attention imbalance as image-centric, while the authors propose a broader “system-mediated” explanation involving functionally redundant system weights.
  • By causally redistributing attention from the system modality toward image and text inputs, the approach significantly reduces the yes-bias and frequently outperforms existing methods.
  • The study also presents evidence that system-mediated attention imbalances can drive over-reliance on coarse input representations—helpful for some tasks but harmful for others—thereby contributing to hallucinations.
  • Overall, the findings establish system attention as a key driver of VLM hallucination and a promising lever for mitigation strategies.

Abstract

Vision-language model (VLM) hallucination is commonly linked to imbalanced allocation of attention across input modalities: system, image and text. However, existing mitigation strategies tend towards an image-centric interpretation of these imbalances, often prioritising increased image attention while giving less consideration to the roles of the other modalities. In this study, we evaluate a more holistic, system-mediated account, which attributes these imbalances to functionally redundant system weights that reduce attention to image and textual inputs. We show that this framework offers a useful empirical perspective on the yes-bias, a common form of hallucination in which VLMs indiscriminately respond `yes'. Causally redistributing attention from the system modality to image and textual inputs substantially suppresses this bias, often outperforming existing approaches. We further present evidence suggesting that system-mediated attention imbalances contribute to the yes-bias by encouraging a default reliance on coarse input representations, which are effective for some tasks but ill-suited to others. Taken together, these findings firmly establish system attention as a key factor in VLM hallucination and highlight its potential as a lever for mitigation.