On-the-Fly VLA Adaptation via Test-Time Reinforcement Learning

arXiv cs.RO / 4/8/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces TT-VLA, a test-time reinforcement learning framework that adapts Vision-Language-Action (VLA) robot policies during inference rather than requiring separate fine-tuning phases or additional data collection.
  • TT-VLA uses a dense reward design based on step-by-step task-progress signals to iteratively improve actions at test time while retaining the original SFT/RL-trained priors.
  • Experiments indicate improved adaptability, stability, and task success for VLAs when facing dynamic, previously unseen scenarios in both simulated and real-world settings.
  • The work positions TT-VLA as a step toward more self-improving, deployment-ready VLAs that can autonomously respond to evolving environments.

Abstract

Vision-Language-Action models have recently emerged as a powerful paradigm for general-purpose robot learning, enabling agents to map visual observations and natural-language instructions into executable robotic actions. Though popular, they are primarily trained via supervised fine-tuning or training-time reinforcement learning, requiring explicit fine-tuning phases, human interventions, or controlled data collection. Consequently, existing methods remain unsuitable for challenging simulated- or physical-world deployments, where robots must respond autonomously and flexibly to evolving environments. To address this limitation, we introduce a Test-Time Reinforcement Learning for VLAs (TT-VLA), a framework that enables on-the-fly policy adaptation during inference. TT-VLA formulates a dense reward mechanism that leverages step-by-step task-progress signals to refine action policies during test time while preserving the SFT/RL-trained priors, making it an effective supplement to current VLA models. Empirical results show that our approach enhances overall adaptability, stability, and task success in dynamic, previously unseen scenarios under simulated and real-world settings. We believe TT-VLA offers a principled step toward self-improving, deployment-ready VLAs.