Grounded World Model for Semantically Generalizable Planning

arXiv cs.RO / 4/14/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a Grounded World Model (GWM) for Model Predictive Control that predicts future outcomes in a vision-language-aligned latent space rather than relying on a goal image distance in a vision-only embedding space.
  • In this GWM-MPC framework, candidate action sequences are scored by the similarity between the predicted future embedding and the task instruction embedding, enabling goal specification via natural language even in new environments.
  • The method is designed to improve semantic generalization compared with vision-language action models that depend on visual-language alignment from pretrained VLMs.
  • On the WISER benchmark (288 test tasks with unseen visual signals and referring expressions), GWM-MPC reports an 87% success rate versus 22% for traditional VLA approaches that strongly overfit the training set.
  • The results are claimed to indicate that grounding world-model planning in a language-aligned latent space can substantially reduce overfitting and improve instruction-driven task performance.

Abstract

In Model Predictive Control (MPC), world models predict the future outcomes of various action proposals, which are then scored to guide the selection of the optimal action. For visuomotor MPC, the score function is a distance metric between a predicted image and a goal image, measured in the latent space of a pretrained vision encoder like DINO and JEPA. However, it is challenging to obtain the goal image in advance of the task execution, particularly in new environments. Additionally, conveying the goal through an image offers limited interactivity compared with natural language. In this work, we propose to learn a Grounded World Model (GWM) in a vision-language-aligned latent space. As a result, each proposed action is scored based on how close its future outcome is to the task instruction, reflected by the similarity of embeddings. This approach transforms the visuomotor MPC to a VLA that surpasses VLM-based VLAs in semantic generalization. On the proposed WISER benchmark, GWM-MPC achieves a 87% success rate on the test set comprising 288 tasks that feature unseen visual signals and referring expressions, yet remain solvable with motions demonstrated during training. In contrast, traditional VLAs achieve an average success rate of 22%, even though they overfit the training set with a 90% success rate.