AI Navigate

Stereo World Model: Camera-Guided Stereo Video Generation

arXiv cs.CV / 3/19/2026

📰 NewsModels & Research

Key Points

  • StereoWorld is a camera-conditioned stereo world model that jointly learns appearance and binocular geometry for end-to-end stereo video generation within the RGB modality, grounding geometry from disparity.
  • It introduces two key designs: a unified camera-frame RoPE for camera-aware positional encoding and a stereo-aware attention decomposition that uses 3D intra-view attention plus horizontal row attention guided by epipolar priors.
  • Across benchmarks, StereoWorld improves stereo consistency, disparity accuracy, and camera-motion fidelity, achieving more than 3x faster generation and about a 5% gain in viewpoint consistency over monocular-then-convert pipelines.
  • Beyond benchmarks, it enables end-to-end binocular VR rendering without depth estimation or inpainting and supports metric-scale depth grounding to aid embodied policy learning.
  • It is compatible with long-video distillation for extended interactive stereo synthesis.

Abstract

We present StereoWorld, a camera-conditioned stereo world model that jointly learns appearance and binocular geometry for end-to-end stereo video generation.Unlike monocular RGB or RGBD approaches, StereoWorld operates exclusively within the RGB modality, while simultaneously grounding geometry directly from disparity. To efficiently achieve consistent stereo generation, our approach introduces two key designs: (1) a unified camera-frame RoPE that augments latent tokens with camera-aware rotary positional encoding, enabling relative, view- and time-consistent conditioning while preserving pretrained video priors via a stable attention initialization; and (2) a stereo-aware attention decomposition that factors full 4D attention into 3D intra-view attention plus horizontal row attention, leveraging the epipolar prior to capture disparity-aligned correspondences with substantially lower compute. Across benchmarks, StereoWorld improves stereo consistency, disparity accuracy, and camera-motion fidelity over strong monocular-then-convert pipelines, achieving more than 3x faster generation with an additional 5% gain in viewpoint consistency. Beyond benchmarks, StereoWorld enables end-to-end binocular VR rendering without depth estimation or inpainting, enhances embodied policy learning through metric-scale depth grounding, and is compatible with long-video distillation for extended interactive stereo synthesis.