Cognitive Pivot Points and Visual Anchoring: Unveiling and Rectifying Hallucinations in Multimodal Reasoning Models

arXiv cs.AI / 4/14/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper identifies a hallucination failure mode in Multimodal Large Reasoning Models called the Reasoning Vision Truth Disconnect (RVTD), where long-chain reasoning errors correlate with cognitive bifurcation points and high-entropy internal states.
  • It argues that the root cause is a breakdown in visual semantic anchoring localized in intermediate network layers, during which the model stops querying visual evidence and instead relies on language priors.
  • The authors propose moving beyond outcome-only supervision by adding fine-grained internal attention guidance to keep reasoning grounded in visual inputs.
  • They introduce V-STAR (Visual Structural Training with Attention Reinforcement), using a Hierarchical Visual Attention Reward (HVAR) within GRPO to dynamically incentivize visual attention at critical high-uncertainty layers.
  • They also present a Forced Reflection Mechanism (FRM) that edits reasoning trajectories by triggering reflection and verification against visual input at the identified high-entropy points to reduce hallucinations.

Abstract

Multimodal Large Reasoning Models (MLRMs) have achieved remarkable strides in visual reasoning through test time compute scaling, yet long chain reasoning remains prone to hallucinations. We identify a concerning phenomenon termed the Reasoning Vision Truth Disconnect (RVTD): hallucinations are strongly correlated with cognitive bifurcation points that often exhibit high entropy states. We attribute this vulnerability to a breakdown in visual semantic anchoring, localized within the network's intermediate layers; specifically, during these high uncertainty transitions, the model fails to query visual evidence, reverting instead to language priors. Consequently, we advocate a shift from solely outcome level supervision to augmenting it with fine grained internal attention guidance. To this end, we propose V-STAR (Visual Structural Training with Attention Reinforcement), a lightweight, holistic training paradigm designed to internalize visually aware reasoning capabilities. Central to our approach is the Hierarchical Visual Attention Reward (HVAR), integrated within the GRPO framework. Upon detecting high entropy states, this mechanism dynamically incentivizes visual attention across critical intermediate layers, thereby anchoring the reasoning process back to the visual input. Furthermore, we introduce the Forced Reflection Mechanism (FRM), a trajectory editing strategy that disrupts cognitive inertia by triggering reflection around high entropy cognitive bifurcation points and encouraging verification of subsequent steps against the visual input, thereby translating external debiasing interventions into an intrinsic capability for hallucination mitigation.