World Action Verifier: Self-Improving World Models via Forward-Inverse Asymmetry

arXiv cs.LG / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes World Action Verifier (WAV), a self-improvement framework for general-purpose world models that can detect and correct its own prediction errors across both optimal and suboptimal actions.
  • WAV factorizes action-conditioned state prediction into two verification targets—state plausibility and action reachability—arguing that these are easier to verify than full state prediction due to data and feature asymmetries.
  • The approach augments a world model with a diverse subgoal generator from video corpora and a sparse inverse model that infers actions from a subset of state features, then enforces cycle consistency across subgoals, inferred actions, and forward rollouts.
  • Experiments on nine tasks across MiniGrid, RoboMimic, and ManiSkill show 2x higher sample efficiency and an 18% improvement in downstream policy performance.
  • The work targets under-explored regimes where existing world-model verification methods struggle, positioning verification as a practical route to robustness and better policy learning.

Abstract

General-purpose world models promise scalable policy evaluation, optimization, and planning, yet achieving the required level of robustness remains challenging. Unlike policy learning, which primarily focuses on optimal actions, a world model must be reliable over a much broader range of suboptimal actions, which are often insufficiently covered by action-labeled interaction data. To address this challenge, we propose World Action Verifier (WAV), a framework that enables world models to identify their own prediction errors and self-improve. The key idea is to decompose action-conditioned state prediction into two factors -- state plausibility and action reachability -- and verify each separately. We show that these verification problems can be substantially easier than predicting future states due to two underlying asymmetries: the broader availability of action-free data and the lower dimensionality of action-relevant features. Leveraging these asymmetries, we augment a world model with (i) a diverse subgoal generator obtained from video corpora and (ii) a sparse inverse model that infers actions from a subset of state features. By enforcing cycle consistency among generated subgoals, inferred actions, and forward rollouts, WAV provides an effective verification mechanism in under-explored regimes, where existing methods typically fail. Across nine tasks spanning MiniGrid, RoboMimic, and ManiSkill, our method achieves 2x higher sample efficiency while improving downstream policy performance by 18%.