CBV: Clean-label Backdoor Attacks on Vision Language Models via Diffusion Models

arXiv cs.AI / 5/5/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper reports that vision-language models (VLMs) are vulnerable to backdoor attacks, and that prior methods often use data poisoning that creates detectable image-text mismatches via visual triggers and altered text labels.
  • It proposes CBV (Clean-Label Backdoor Attack on VLMs via Diffusion Models), which generates natural-looking poisoned samples using diffusion models by steering the reverse diffusion process with modified scores via score matching.
  • CBV further improves attack effectiveness by using multimodal guidance that incorporates text information derived from the triggered images during generation.
  • To increase stealth, the method introduces a GradCAM-guided mask (GM) so that perturbations are applied only to the most semantically important regions rather than the whole image.
  • Experiments on MSCOCO and VQA v2 using four representative VLMs show over 80% attack success rate (ASR) while keeping normal model functionality largely intact.

Abstract

Vision-Language Models (VLMs) have achieved remarkable success in tasks such as image captioning and visual question answering (VQA). However, as their applications become increasingly widespread, recent studies have revealed that VLMs are vulnerable to backdoor attacks. Existing backdoor attacks on VLMs primarily rely on data poisoning by adding visual triggers and modifying text labels, where the induced image-text mismatch makes poisoned samples easy to detect. To address this limitation, we propose the Clean-Label Backdoor Attack on VLMs via Diffusion Models (CBV), which leverages diffusion models to generate natural poisoned examples via score matching. Specifically, CBV modifies the score during the reverse generation process of the diffusion model to guide the generation of poisoned samples that contain triggered image features. To further enhance the effectiveness of the attack, we incorporate the textual information of the triggered images as multimodal guidance during generation. Moreover, to enhance stealthiness, we introduce a GradCAM-guided Mask (GM) that restricts modifications to only the most semantically important regions, rather than the entire image. We evaluate our method on MSCOCO and VQA v2 with four representative VLMs, achieving over 80% ASR while preserving normal functionality.