Fighting Hallucinations with Counterfactuals: Diffusion-Guided Perturbations for LVLM Hallucination Suppression
arXiv cs.CV / 3/12/2026
📰 NewsModels & Research
Key Points
- Introduces CIPHER, a training-free method that suppresses vision-induced hallucinations in LVLMs by feature-level correction using counterfactual visual perturbations.
- Builds OHC-25K, a counterfactual dataset of diffusion-edited images paired with their original captions to reveal systematic hallucination-related shifts.
- Inference-time, CIPHER projects intermediate hidden states away from a low-rank subspace associated with vision-induced hallucinations, reducing hallucinations while preserving task performance.
- Empirical results across multiple benchmarks show significant reductions in hallucination rates with maintained or improved task effectiveness, and code is released for reproducibility.
Related Articles

ラピダス、半導体設計AIエージェント「国内2社海外1社が使用中」
日経XTECH

Superposition and the Capsule: Quantum State Collapse Meets AI Identity
Dev.to

The Basilisk Inversion: Why Coercive AI Futures Are Thermodynamically Unlikely
Dev.to

The Loop as Laboratory: What 3,190 Cycles of Autonomous AI Operation Reveal
Dev.to

MiMo-V2-Pro & Omni & TTS: "We will open-source — when the models are stable enough to deserve it."
Reddit r/LocalLLaMA