RL Token: Bootstrapping Online RL with Vision-Language-Action Models

arXiv cs.LG / 4/28/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper proposes RL Token (RLT), a lightweight method for sample-efficient online reinforcement learning fine-tuning of pretrained vision-language-action (VLA) models using only a few hours of real-world practice.
  • RLT modifies a VLA to expose an “RL token” that retains task-relevant pretrained knowledge while providing an efficient interface for online RL, and it trains a small actor-critic head on top of this token to refine actions.
  • The learned policy is anchored to the underlying VLA to preserve pretrained capabilities while improving precision and responsiveness needed for real-world manipulation.
  • Experiments on four real-robot tasks (screw installation, zip tie fastening, charger insertion, and Ethernet insertion) show up to 3× speed improvements on the most difficult task phase and substantially higher success rates within minutes to a few hours.
  • In some tasks, RLT can even outperform the speed of human teleoperation, highlighting its potential for fast and practical robotic skill adaptation.

Abstract

Vision-language-action (VLA) models can learn to perform diverse manipulation skills "out of the box," but achieving the precision and speed that real-world tasks demand requires further fine-tuning -- for example, via reinforcement learning (RL). We introduce a lightweight method that enables sample-efficient online RL fine-tuning of pretrained VLAs using just a few hours of real-world practice. We (1) adapt the VLA to expose an "RL token," a compact readout representation that preserves task-relevant pretrained knowledge while serving as an efficient interface for online RL, and (2) train a small actor-critic head on this RL token to refine the actions, while anchoring the learned policy to the VLA. Online RL with the RL token (RLT) makes it possible to fine-tune even large VLAs with RL quickly and efficiently. Across four real-robot tasks (screw installation, zip tie fastening, charger insertion, and Ethernet insertion), RLT improves the speed on the hardest part of the task by up to 3x and raises success rates significantly within minutes to a few hours of practice. It can even surpass the speed of human teleoperation on some of the tasks.