Fighting Hallucinations with Counterfactuals: Diffusion-Guided Perturbations for LVLM Hallucination Suppression
arXiv cs.CV / 3/12/2026
📰 NewsModels & Research
Key Points
- Introduces CIPHER, a training-free method that suppresses vision-induced hallucinations in LVLMs by feature-level correction using counterfactual visual perturbations.
- Builds OHC-25K, a counterfactual dataset of diffusion-edited images paired with their original captions to reveal systematic hallucination-related shifts.
- Inference-time, CIPHER projects intermediate hidden states away from a low-rank subspace associated with vision-induced hallucinations, reducing hallucinations while preserving task performance.
- Empirical results across multiple benchmarks show significant reductions in hallucination rates with maintained or improved task effectiveness, and code is released for reproducibility.
Related Articles
[D] Matryoshka Representation Learning
Reddit r/MachineLearning
Two new Qwen3.5 “Neo” fine‑tunes focused on fast, efficient reasoning
Reddit r/LocalLLaMA

HKIC, Gobi Partners and HKU team up for fund backing university research start-ups
SCMP Tech
Yann LeCun’s New LeWorldModel (LeWM) Research Targets JEPA Collapse in Pixel-Based Predictive World Modeling
MarkTechPost
Streaming experts
Simon Willison's Blog