Understanding Adversarial Transferability in Vision-Language Models for Autonomous Driving: A Cross-Architecture Analysis

arXiv cs.CV / 5/1/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The study investigates how robust vision-language models (VLMs) used in autonomous driving are to physical adversarial attacks, focusing on whether attacks can transfer across different model architectures.
  • Researchers conduct a cross-architecture evaluation using three representative VLM-based driving architectures (Dolphins, OmniDrive, and LeapVAD) with physically realizable patch attacks on roadside infrastructure in crosswalk and highway scenarios.
  • The results show high cross-architecture transferability, with reported transfer rates of 73–91% and mean transfer metrics (TR) of 0.815 for crosswalk and 0.833 for highway.
  • Frame-level manipulation persists for a large portion of the critical decision window (64.7–79.4%) even when adversarial patches are not optimized for the target model, suggesting a practical security risk.
  • Overall, the findings indicate that attackers may not need knowledge of the specific deployed VLM architecture to induce harmful perception or decision disruptions in driving contexts.

Abstract

Vision-language models (VLMs) are increasingly used in autonomous driving because they combine visual perception with language-based reasoning, supporting more interpretable decision-making, yet their robustness to physical adversarial attacks, especially whether such attacks transfer across different VLM architectures, is not well understood and poses a practical risk when attackers do not know which model a vehicle uses. We address this gap with a systematic cross-architecture study of adversarial transferability in VLM-based driving, evaluating three representative architectures (Dolphins, OmniDrive, and LeapVAD) using physically realizable patches placed on roadside infrastructure in both crosswalk and highway scenarios. Our transfer-matrix evaluation shows high cross-architecture effectiveness, with transfer rates of 73-91% (mean TR = 0.815 for crosswalk and 0.833 for highway) and sustained frame-level manipulation over 64.7-79.4% of the critical decision window even when patches are not optimized for the target model.