Uni-World VLA: Interleaved World Modeling and Planning for Autonomous Driving

arXiv cs.RO / 3/31/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Uni-World VLA, a unified vision-language-action model for autonomous driving that interleaves future frame prediction with trajectory planning rather than running them in separate open-loop stages.
  • By alternating step-by-step imagination of future observations and ego actions, the method keeps planning continuously conditioned on the evolving predicted scenes, forming a closed-loop between world modeling and control.
  • It further improves long-horizon scene prediction by integrating monocular depth cues into the frame representations to strengthen geometric understanding.
  • Experiments on the NAVSIM benchmark report competitive closed-loop planning performance alongside high-fidelity future frame predictions, suggesting tighter coupling of prediction and planning can improve adaptive driving in dynamic traffic.

Abstract

Autonomous driving requires reasoning about how the environment evolves and planning actions accordingly. Existing world-model-based approaches typically predict future scenes first and plan afterwards, resulting in open-loop imagination that may drift from the actual decision process. In this paper, we present Uni-World VLA, a unified vision-language-action (VLA) model that tightly interleaves future frame prediction and trajectory planning. Instead of generating a full world rollout before planning, our model alternates between predicting future frames and ego actions step by step, allowing planning decisions to be continuously conditioned on the imagined future observations. This interleaved generation forms a closed-loop interaction between world modeling and control, enabling more adaptive decision-making in dynamic traffic scenarios. In addition, we incorporate monocular depth information into frames to provide stronger geometric cues for world modeling, improving long-horizon scene prediction. Experiments on the NAVSIM benchmark show that our approach achieves competitive closed-loop planning performance while producing high-fidelity future frame predictions. These results demonstrate that tightly coupling world prediction and planning is a promising direction for scalable VLA driving systems.