Revealing Physical-World Semantic Vulnerabilities: Universal Adversarial Patches for Infrared Vision-Language Models
arXiv cs.CV / 4/6/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper studies how infrared vision-language models (IR-VLMs), despite being promising for low-visibility perception, remain vulnerable to physical-world adversarial attacks that are not well addressed by prior RGB-focused methods.
- It introduces Universal Curved-Grid Patch (UCGP), a deployable universal adversarial patch framework that uses curved-grid mesh parameterization and a representation-level objective (e.g., subspace departure, topology disruption, and stealth) rather than changing prompts or labels.
- The method further improves real-world robustness under domain shift by combining Meta Differential Evolution with EOT-augmented TPS deformation modeling to better simulate physical transformations.
- Experiments show UCGP reliably degrades semantic understanding across multiple IR-VLM architectures, with strong transferability across models and datasets and demonstrable effectiveness in real physical settings.
- Overall, the work highlights a previously underappreciated robustness weakness in infrared multimodal systems and suggests existing defenses may not cover representation-space disruption threats.
Related Articles

Black Hat Asia
AI Business

Оказывается, эта нейросеть рисует бесплатно. Я узнал случайно.
Dev.to

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Three-Layer Memory Governance: Core, Provisional, Private
Dev.to

I Researched AI Prompting So You Don’t Have To
Dev.to