Repurposing 3D Generative Model for Autoregressive Layout Generation

arXiv cs.CV / 4/20/2026

📰 NewsDeveloper Stack & InfrastructureTools & Practical UsageModels & Research

Key Points

  • The article introduces LaviGen, a framework that repurposes 3D generative models to generate 3D layouts directly in native 3D space.
  • Instead of deriving layouts from text via separate inference steps, LaviGen formulates layout generation as an autoregressive process that explicitly models geometric relationships and physical constraints.
  • It also proposes an adapted 3D diffusion model that fuses scene, object, and instruction information, and uses dual-guidance self-rollout distillation to improve both efficiency and spatial accuracy.
  • Experiments on the LayoutVLM benchmark report 19% higher physical plausibility than prior state of the art and 65% faster computation, and the authors release code on GitHub.

Abstract

We introduce LaviGen, a framework that repurposes 3D generative models for 3D layout generation. Unlike previous methods that infer object layouts from textual descriptions, LaviGen operates directly in the native 3D space, formulating layout generation as an autoregressive process that explicitly models geometric relations and physical constraints among objects, producing coherent and physically plausible 3D scenes. To further enhance this process, we propose an adapted 3D diffusion model that integrates scene, object, and instruction information and employs a dual-guidance self-rollout distillation mechanism to improve efficiency and spatial accuracy. Extensive experiments on the LayoutVLM benchmark show LaviGen achieves superior 3D layout generation performance, with 19% higher physical plausibility than the state of the art and 65% faster computation. Our code is publicly available at https://github.com/fenghora/LaviGen.