EVA: Aligning Video World Models with Executable Robot Actions via Inverse Dynamics Rewards

arXiv cs.RO / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper identifies an “executability gap” in video-based world models for robotics: visually plausible rollouts can still generate robot actions that violate rigid-body/kinematic constraints when decoded by an inverse dynamics model.
  • It proposes Executable Video Alignment (EVA), a reinforcement-learning post-training framework that uses an inverse dynamics model trained on real robot trajectories as a reward model for evaluating generated videos.
  • EVA incentivizes smoother, physically consistent motion (based on velocity, acceleration, and jerk) while penalizing actions that break embodiment constraints, improving alignment between visual prediction and feasible robot control.
  • The authors report that the reward signal remains useful even with severe visual artifacts, because those artifacts often induce unstable or out-of-bounds action sequences.
  • Experiments on the RoboTwin benchmark and a real bimanual robot show EVA reduces embodiment-specific artifacts in rollouts and improves task execution success.

Abstract

Video generative models are increasingly used as world models for robotics, where a model generates a future visual rollout conditioned on the current observation and task instruction, and an inverse dynamics model (IDM) converts the generated frames into executable robot actions. However, current video world models lack explicit executability constraints. As a result, visually coherent rollouts may still violate rigid-body and kinematic consistency, producing unstable or infeasible control commands when decoded by an IDM. We refer to this mismatch between visual generation and physically executable control as the executability gap. While this gap can be mitigated at inference time using techniques such as rejection sampling, such approaches are inefficient due to the high cost of video generation. In this paper, we leverage the executability gap as a training signal and introduce Executable Video Alignment (EVA), a reinforcement-learning post-training framework for aligning video world models. EVA trains an inverse dynamics model on real robot trajectories and repurposes it as a reward model that evaluates generated videos through the action sequences they induce, encouraging smooth motions measured by velocity, acceleration, and jerk while penalizing actions that violate embodiment constraints. Importantly, the reward remains informative even when generated videos contain severe visual artifacts, since such artifacts typically translate into unstable or out-of-bound actions. Experiments on the RoboTwin benchmark and a real bimanual robot show that EVA reduces embodiment-specific artifacts in generated rollouts and improves downstream task execution success.