Understanding and Mitigating Hallucinations in Multimodal Chain-of-Thought Models

arXiv cs.CV / 3/31/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • Multimodal Chain-of-Thought (MCoT) models show strong performance on complex visual reasoning but suffer from severe hallucinations, partly tied to degraded visual attention during generation.
  • The study tests whether MCoT hallucinations have unique underlying causes and finds that fabricated text mainly emerges during “associative reasoning” steps referred to as divergent thinking.
  • It proposes a simple decoding-time strategy to localize the divergent-thinking steps and intervene to reduce hallucinations.
  • Experimental results indicate the new method significantly outperforms prior hallucination mitigation approaches.
  • The approach is designed to be modular, allowing easy integration with other hallucination mitigation techniques for additional gains, with code released publicly.

Abstract

Multimodal Chain-of-Thought (MCoT) models have demonstrated impressive capability in complex visual reasoning tasks. Unfortunately, recent studies reveal that they suffer from severe hallucination problems due to diminished visual attention during the generation process. However, visual attention decay is a well-studied problem in Large Vision-Language Models (LVLMs). Considering the fundamental differences in reasoning processes between MCoT models and traditional LVLMs, we raise a basic question: Whether MCoT models have unique causes of hallucinations? To answer this question, we systematically investigate the hallucination patterns of MCoT models and find that fabricated texts are primarily generated in associative reasoning steps, which we term divergent thinking. Leveraging these insights, we introduce a simple yet effective strategy that can effectively localize divergent thinking steps and intervene in the decoding process to mitigate hallucinations. Extensive experiments show that our method outperforms existing methods by a large margin. More importantly, our proposed method can be conveniently integrated with other hallucination mitigation methods and further boost their performance. The code is publicly available at https://github.com/ASGO-MM/MCoT-hallucination.