UniLACT: Depth-Aware RGB Latent Action Learning for Vision-Language-Action Models

arXiv cs.RO / 4/10/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes UniLACT, a depth-aware transformer-based vision-language-action (VLA) model that improves latent action pretraining by incorporating 3D geometric structure instead of relying on RGB appearance alone.
  • It introduces UniLARN, a unified latent action learning framework that uses inverse and forward dynamics objectives to learn a shared embedding space for RGB and depth while explicitly modeling cross-modal interactions.
  • The learned modality-specific and unified latent action representations are used as pseudo-labels to enable depth-aware pretraining, giving downstream policies stronger spatial priors for contact-rich manipulation.
  • Experiments in both simulation and real-world settings show that UniLACT outperforms RGB-based latent action baselines across in-domain and out-of-domain pretraining, including both seen and unseen tasks.

Abstract

Latent action representations learned from unlabeled videos have recently emerged as a promising paradigm for pretraining vision-language-action (VLA) models without explicit robot action supervision. However, latent actions derived solely from RGB observations primarily encode appearance-driven dynamics and lack explicit 3D geometric structure, which is essential for precise and contact-rich manipulation. To address this limitation, we introduce UniLACT, a transformer-based VLA model that incorporates geometric structure through depth-aware latent pretraining, enabling downstream policies to inherit stronger spatial priors. To facilitate this process, we propose UniLARN, a unified latent action learning framework based on inverse and forward dynamics objectives that learns a shared embedding space for RGB and depth while explicitly modeling their cross-modal interactions. This formulation produces modality-specific and unified latent action representations that serve as pseudo-labels for the depth-aware pretraining of UniLACT. Extensive experiments in both simulation and real-world settings demonstrate the effectiveness of depth-aware unified latent action representations. UniLACT consistently outperforms RGB-based latent action baselines under in-domain and out-of-domain pretraining regimes, as well as on both seen and unseen manipulation tasks.The project page is at https://manishgovind.github.io/unilact-vla/