FlowHijack: A Dynamics-Aware Backdoor Attack on Flow-Matching Vision-Language-Action Models

arXiv cs.CV / 4/14/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper identifies a previously unexplored backdoor vulnerability in flow-matching Vision-Language-Action (VLA) models, specifically in their continuous vector-field (dynamics) action generation mechanism.
  • It introduces “FlowHijack,” the first framework designed to target the underlying dynamics rather than relying on attacks for autoregressive, discretized VLA models.
  • FlowHijack uses a τ-conditioned injection strategy to manipulate the early phase of action generation and a dynamics-mimicry regularizer to keep malicious behavior aligned with normal motion patterns.
  • Experiments report high attack success with stealthy, context-aware triggers, while preserving benign task performance and producing malicious actions that are behaviorally indistinguishable from legitimate ones.
  • The results suggest continuous embodied models have a significant security risk and motivate the development of defenses focused on internal generative dynamics.

Abstract

Vision-Language-Action (VLA) models are emerging as a cornerstone for robotics, with flow-matching policies like \pi_0 showing great promise in generating smooth, continuous actions. As these models advance, their unique action generation mechanism - the vector field dynamics - presents a critical yet unexplored security vulnerability, particularly backdoor vulnerabilities. Existing backdoor attacks designed for autoregressive discretization VLAs cannot be directly applied to this new continuous dynamics. We introduce FlowHijack, the first backdoor attack framework to systematically target the underlying vector-field dynamics of flow-matching VLAs. Our method combines a novel \tau-conditioned injection strategy, which manipulates the initial phase of the action generation, with a dynamics mimicry regularizer. Experiments demonstrate that FlowHijack achieves high attack success rates using stealthy, context-aware triggers where prior works failed. Crucially, it preserves benign task performance and, by enforcing kinematic similarity, generates malicious actions that are behaviorally indistinguishable from normal actions. Our findings reveal a significant vulnerability in continuous embodied models, highlighting the urgent need for defenses targeting the model's internal generative dynamics.