How Vulnerable Is My Learned Policy? Universal Adversarial Perturbation Attacks On Modern Behavior Cloning Policies
arXiv cs.RO / 4/27/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper presents the first systematic study of universal adversarial perturbation attacks against a range of modern imitation learning and behavior cloning algorithms, including Vanilla BC, IBC, Diffusion Policy, and VQ-BET.
- It evaluates vulnerability under multiple threat models—white-box, grey-box, and black-box—showing that adversarial perturbations can reliably degrade learned policies.
- Experimental results indicate that most existing methods are highly vulnerable, including black-box transfer attacks where adversarial examples generated for one algorithm can succeed against others.
- The authors make cross-algorithm comparisons for both white-box and black-box settings and provide links to videos and code to support further research.
- Overall, the study surfaces a key security limitation of modern imitation learning and motivates future work to mitigate these weaknesses.
Related Articles

Subagents: The Building Block of Agentic AI
Dev.to

DeepSeek-V4 Models Could Change Global AI Race
AI Business

Got OpenAI's privacy filter model running on-device via ExecuTorch
Reddit r/LocalLLaMA

The Agent-Skill Illusion: Why Prompt-Based Control Fails in Multi-Agent Business Consulting Systems
Dev.to

We Built a Voice AI Receptionist in 8 Weeks — Every Decision We Made and Why
Dev.to