Toward Consistent World Models with Multi-Token Prediction and Latent Semantic Enhancement

arXiv cs.LG / 4/8/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper examines whether large language models can form coherent internal world models, arguing that multi-token prediction (MTP) can push representations toward internally consistent “belief states.”
  • It provides a theoretical analysis of MTP’s gradient inductive bias, claiming MTP induces representational contractivity through gradient coupling that supports convergence.
  • The authors identify a failure mode of standard MTP: “structural hallucinations,” where discrete token supervision leads to illegal latent-space shortcuts that break environmental constraints.
  • To mitigate this, they introduce Latent Semantic Enhancement MTP (LSE-MTP), which anchors prediction targets to ground-truth hidden-state trajectories to better connect token-level outputs with continuous latent dynamics.
  • Experiments on synthetic graphs and the Manhattan Taxi Ride domain show LSE-MTP improves representation alignment, reduces structural hallucinations, and increases robustness under perturbations.

Abstract

Whether Large Language Models (LLMs) develop coherent internal world models remains a core debate. While conventional Next-Token Prediction (NTP) focuses on one-step-ahead supervision, Multi-Token Prediction (MTP) has shown promise in learning more structured representations. In this work, we provide a theoretical perspective analyzing the gradient inductive bias of MTP, supported by empirical evidence, showing that MTP promotes the convergence toward internal belief states by inducing representational contractivity via gradient coupling. However, we reveal that standard MTP often suffers from structural hallucinations, where discrete token supervision encourages illegal shortcuts in latent space that violate environmental constraints. To address this, we propose a novel method Latent Semantic Enhancement MTP (LSE-MTP), which anchors predictions to ground-truth hidden state trajectories. Experiments on synthetic graphs and real-world Manhattan Taxi Ride show that LSE-MTP effectively bridges the gap between discrete tokens and continuous state representations, enhancing representation alignment, reducing structural hallucinations, and improving robustness to perturbations.