WorldMesh: Generating Navigable Multi-Room 3D Scenes via Mesh-Conditioned Image Diffusion

arXiv cs.CV / 3/25/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes WorldMesh, a geometry-first method for generating large, navigable multi-room 3D scenes that addresses text-to-image/video consistency limits caused by missing explicit geometry.
  • It decouples scene generation into two stages: creating a mesh “scaffold” that captures environment structure (e.g., walls and floors), and then synthesizing realistic appearance conditioned on that mesh.
  • Starting from a text description, the system constructs a geometry mesh, then uses image synthesis plus segmentation and object reconstruction to place objects with coherent layouts on the scaffold.
  • By rendering the mesh scaffold to condition subsequent image synthesis, the approach aims to provide a structural backbone that improves object/scene-level consistency while scaling to arbitrarily sized, highly populated environments.
  • The authors position the work as a meaningful step toward generating environment-scale, immersive 3D worlds with both robust 3D consistency and photorealistic detail.

Abstract

Recent progress in image and video synthesis has inspired their use in advancing 3D scene generation. However, we observe that text-to-image and -video approaches struggle to maintain scene- and object-level consistency beyond a limited environment scale due to the absence of explicit geometry. We thus present a geometry-first approach that decouples this complex problem of large-scale 3D scene synthesis into its structural composition, represented as a mesh scaffold, and realistic appearance synthesis, which leverages powerful image synthesis models conditioned on the mesh scaffold. From an input text description, we first construct a mesh capturing the environment's geometry (walls, floors, etc.), and then use image synthesis, segmentation and object reconstruction to populate the mesh structure with objects in realistic layouts. This mesh scaffold is then rendered to condition image synthesis, providing a structural backbone for consistent appearance generation. This enables scalable, arbitrarily-sized 3D scenes of high object richness and diversity, combining robust 3D consistency with photorealistic detail. We believe this marks a significant step toward generating truly environment-scale, immersive 3D worlds.