PDA: Text-Augmented Defense Framework for Robust Vision-Language Models against Adversarial Image Attacks
arXiv cs.CV / 4/2/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces Paraphrase-Decomposition-Aggregation (PDA), a training-free inference-time defense intended to make vision-language models (VLMs) more robust to adversarial image perturbations.
- PDA improves robustness by applying test-time prompt paraphrasing, decomposing questions, and aggregating consistency across the augmented text inputs, without changing the underlying VLM.
- To manage the compute/latency trade-off, the authors instantiate PDA as “invariants” that reduce inference cost while preserving most of the robustness improvements.
- Experiments across multiple VLM architectures and benchmarks for visual question answering, classification, and captioning report consistent robustness gains against diverse adversarial attacks while maintaining competitive accuracy on clean (non-adversarial) inputs.
Related Articles

Benchmarking Batch Deep Reinforcement Learning Algorithms
Dev.to

Qwen3.6-Plus: Alibaba's Quiet Giant in the AI Race Delivers a Million-Token Enterprise Powerhouse
Dev.to

How To Leverage AI for Back-Office Headcount Optimization
Dev.to
Is 1-bit and TurboQuant the future of OSS? A simulation for Qwen3.5 models.
Reddit r/LocalLLaMA
SOTA Language Models Under 14B?
Reddit r/LocalLLaMA