Language-Conditioned World Modeling for Visual Navigation

arXiv cs.CV / 3/31/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies language-conditioned visual navigation, where an embodied agent must follow natural language instructions using only an initial egocentric observation and no goal images, making language grounding central to the control problem.
  • It introduces the LCVN Dataset, containing 39,016 trajectories and 117,048 human-verified instructions across multiple environments and instruction styles to support reproducible benchmarking.
  • The authors frame the task as language-conditioned open-loop trajectory prediction and propose two model families that connect language grounding, future-state (imagination) prediction, and action generation.
  • One approach (LCVN-WM + LCVN-AC) uses a diffusion-based world model with an actor-critic policy operating in the model’s latent space, yielding more temporally coherent rollouts.
  • The other approach (LCVN-Uni) uses an autoregressive multimodal architecture to predict future observations and actions, showing better generalization to unseen environments.

Abstract

We study language-conditioned visual navigation (LCVN), in which an embodied agent is asked to follow a natural language instruction based only on an initial egocentric observation. Without access to goal images, the agent must rely on language to shape its perception and continuous control, making the grounding problem particularly challenging. We formulate this problem as open-loop trajectory prediction conditioned on linguistic instructions and introduce the LCVN Dataset, a benchmark of 39,016 trajectories and 117,048 human-verified instructions that supports reproducible research across a range of environments and instruction styles. Using this dataset, we develop LCVN frameworks that link language grounding, future-state prediction, and action generation through two complementary model families. The first family combines LCVN-WM, a diffusion-based world model, with LCVN-AC, an actor-critic agent trained in the latent space of the world model. The second family, LCVN-Uni, adopts an autoregressive multimodal architecture that predicts both actions and future observations. Experiments show that these families offer different advantages: the former provides more temporally coherent rollouts, whereas the latter generalizes better to unseen environments. Taken together, these observations point to the value of jointly studying language grounding, imagination, and policy learning in a unified task setting, and LCVN provides a concrete basis for further investigation of language-conditioned world models. The code is available at https://github.com/F1y1113/LCVN.