Lost in Diffusion: Uncovering Hallucination Patterns and Failure Modes in Diffusion Large Language Models
arXiv cs.CL / 4/14/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper presents a first controlled comparative study of hallucination patterns in diffusion large language models (dLLMs) versus autoregressive (AR) models under matched architecture, scale, and pre-training weights.
- It finds that current dLLMs hallucinate more frequently than AR counterparts, indicating weaker faithfulness despite progress on general tasks.
- Analysis of inference-time compute shows different generation dynamics: quasi-autoregressive decoding saturates early, while non-sequential decoding can enable continuous refinement.
- The study identifies diffusion-specific hallucination failure modes such as premature termination, incomplete denoising, and context intrusion, highlighting reliability risks unique to the diffusion process.
- The authors release accompanying code at the provided GitHub repository to support further investigation and replication of their evaluation approach.
Related Articles

Black Hat Asia
AI Business

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Don't forget, there is more than forgetting: new metrics for Continual Learning
Dev.to

Microsoft MAI-Image-2-Efficient Review 2026: The AI Image Model Built for Production Scale
Dev.to
Bit of a strange question?
Reddit r/artificial