VLA-OPD: Bridging Offline SFT and Online RL for Vision-Language-Action Models via On-Policy Distillation

arXiv cs.RO / 3/30/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces VLA-OPD, a post-training framework for vision-language-action (VLA) robotic models that combines the efficiency of offline supervised fine-tuning (SFT) with the robustness of online reinforcement learning (RL).
  • Instead of using sparse environmental rewards, VLA-OPD uses an expert teacher to provide dense, token-level supervision on the student’s own self-generated trajectories, enabling corrective learning on policy-induced states.
  • The method uses a Reverse-KL objective to stabilize learning, aiming to avoid the entropy issues of Forward-KL and the premature entropy collapse associated with Hard cross-entropy.
  • Experiments on LIBERO and RoboTwin2.0 show that VLA-OPD improves sample efficiency versus RL, increases robustness versus SFT, and mitigates catastrophic forgetting of pre-trained capabilities.
  • Overall, the approach frames post-training as “gentle alignment” that preserves prior generalization while correcting errors during distribution shift induced by the evolving policy.

Abstract

Although pre-trained Vision-Language-Action (VLA) models exhibit impressive generalization in robotic manipulation, post-training remains crucial to ensure reliable performance during deployment. However, standard offline Supervised Fine-Tuning (SFT) suffers from distribution shifts and catastrophic forgetting of pre-trained capabilities, while online Reinforcement Learning (RL) struggles with sparse rewards and poor sample efficiency. In this paper, we propose On-Policy VLA Distillation (VLA-OPD), a framework bridging the efficiency of SFT with the robustness of RL. Instead of relying on sparse environmental rewards, VLA-OPD leverages an expert teacher to provide dense, token-level supervision on the student's self-generated trajectories. This enables active error correction on policy-induced states while preserving pre-trained general capabilities through gentle alignment. Crucially, we formulate VLA-OPD via a Reverse-KL objective. Unlike standard Forward-KL that induces mode-covering entropy explosion, or Hard-CE that causes premature entropy collapse, our bounded mode-seeking objective ensures stable policy learning by filtering out the teacher's epistemic uncertainty while maintaining action diversity. Experiments on LIBERO and RoboTwin2.0 benchmarks demonstrate that VLA-OPD significantly improves sample efficiency over RL and robustness over SFT, while effectively mitigating catastrophic forgetting during post-training.