RedVLA: Physical Red Teaming for Vision-Language-Action Models

arXiv cs.RO / 4/27/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces RedVLA, the first dedicated red-teaming framework aimed at detecting physical safety risks before deploying Vision-Language-Action (VLA) models in the real world.
  • RedVLA uses a two-stage pipeline: Risk Scenario Synthesis to create task-feasible initial risk scenes that entangle the risk factor with the model’s execution, and Risk Amplification to reliably elicit unsafe behaviors across different VLA models.
  • Risk Amplification is performed using gradient-free optimization iteratively refined by trajectory features, improving stability when testing heterogeneous models.
  • Experiments across six representative VLA models show RedVLA can discover diverse unsafe behaviors and reach an attack success rate (ASR) of up to 95.5% within 10 optimization iterations.
  • The authors also propose SimpleVLA-Guard, a lightweight safety guard trained using data generated by RedVLA, and release the data, assets, and code publicly.

Abstract

The real-world deployment of Vision-Language-Action (VLA) models remains limited by the risk of unpredictable and irreversible physical harm. However, we currently lack effective mechanisms to proactively detect these physical safety risks before deployment. To address this gap, we propose \textbf{RedVLA}, the first red teaming framework for physical safety in VLA models. We systematically uncover unsafe behaviors through a two-stage process: (I) \textbf{Risk Scenario Synthesis} constructs a valid and task-feasible initial risk scene. Specifically, it identifies critical interaction regions from benign trajectories and positions the risk factor within these regions, aiming to entangle it with the VLA's execution flow and elicit a target unsafe behavior. (II) \textbf{Risk Amplification} ensures stable elicitation across heterogeneous models. It iteratively refines the risk factor state through gradient-free optimization guided by trajectory features. Experiments on six representative VLA models show that RedVLA uncovers diverse unsafe behaviors and achieves the ASR up to 95.5\% within 10 optimization iterations. To mitigate these risks, we further propose SimpleVLA-Guard, a lightweight safety guard built from RedVLA-generated data. Our data, assets, and code are available \href{https://redvla.github.io}{here}.