Decoding by Perturbation: Mitigating MLLM Hallucinations via Dynamic Textual Perturbation
arXiv cs.CL / 4/15/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that multimodal LLM hallucinations often arise because language priors over-dominate visual evidence during decoding, especially in how visual grounding reacts to textual phrasing.
- It proposes “Decoding by Perturbation (DeP),” a training-free method that introduces controlled, multi-level textual perturbations during inference to elicit and manage latent language priors.
- DeP uses attention variance to reinforce stable, evidence-aligned regions in the feature space while suppressing suspicious noise.
- It further estimates an interpretable “prior drift direction” from logits statistics to counteract probability biases caused by textual co-occurrences.
- Experiments reportedly show DeP reduces hallucinations and improves benchmark performance across multiple evaluations.




