Lost in Diffusion: Uncovering Hallucination Patterns and Failure Modes in Diffusion Large Language Models

arXiv cs.CL / 4/14/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper presents a first controlled comparative study of hallucination patterns in diffusion large language models (dLLMs) versus autoregressive (AR) models under matched architecture, scale, and pre-training weights.
  • It finds that current dLLMs hallucinate more frequently than AR counterparts, indicating weaker faithfulness despite progress on general tasks.
  • Analysis of inference-time compute shows different generation dynamics: quasi-autoregressive decoding saturates early, while non-sequential decoding can enable continuous refinement.
  • The study identifies diffusion-specific hallucination failure modes such as premature termination, incomplete denoising, and context intrusion, highlighting reliability risks unique to the diffusion process.
  • The authors release accompanying code at the provided GitHub repository to support further investigation and replication of their evaluation approach.

Abstract

While Diffusion Large Language Models (dLLMs) have emerged as a promising non-autoregressive paradigm comparable to autoregressive (AR) models, their faithfulness, specifically regarding hallucination, remains largely underexplored. To bridge this gap, we present the first controlled comparative study to evaluate hallucination patterns in dLLMs. Our results demonstrate that current dLLMs exhibit a higher propensity for hallucination than AR counterparts controlled for architecture, scale, and pre-training weights. Furthermore, an analysis of inference-time compute reveals divergent dynamics: while quasi-autoregressive generation suffers from early saturation, non-sequential decoding unlocks potential for continuous refinement. Finally, we identify distinct failure modes unique to the diffusion process, including premature termination, incomplete denoising, and context intrusion. Our findings underscore that although dLLMs have narrowed the performance gap on general tasks, their distinct hallucination mechanisms pose a critical challenge to model reliability. Our code is available at https://github.com/ZeroLoss-Lab/Lost-in-Diffusion