Human Interaction-Aware 3D Reconstruction from a Single Image

arXiv cs.CV / 4/8/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses a key gap in single-image textured 3D human reconstruction by targeting multi-person scenes where existing single-individual methods produce overlaps, missing occluded geometry, and distorted interactions.
  • It proposes HUG3D, which jointly models group-level context and instance-level details using a canonical orthographic transformation to reduce perspective-induced geometric distortion.
  • The HUG-MVD diffusion component generates complete multi-view normals and images while resolving occlusions and proximity via coordinated group/individual modeling.
  • The HUG-GR geometry module refines the 3D structure by optimizing with physics-based interaction priors to enforce physical plausibility and accurately represent inter-human contact.
  • Experiments report that HUG3D significantly outperforms prior single-human and multi-human approaches, yielding physically plausible, high-fidelity reconstructions from one image.

Abstract

Reconstructing textured 3D human models from a single image is fundamental for AR/VR and digital human applications. However, existing methods mostly focus on single individuals and thus fail in multi-human scenes, where naive composition of individual reconstructions often leads to artifacts such as unrealistic overlaps, missing geometry in occluded regions, and distorted interactions. These limitations highlight the need for approaches that incorporate group-level context and interaction priors. We introduce a holistic method that explicitly models both group- and instance-level information. To mitigate perspective-induced geometric distortions, we first transform the input into a canonical orthographic space. Our primary component, Human Group-Instance Multi-View Diffusion (HUG-MVD), then generates complete multi-view normals and images by jointly modeling individuals and group context to resolve occlusions and proximity. Subsequently, the Human Group-Instance Geometric Reconstruction (HUG-GR) module optimizes the geometry by leveraging explicit, physics-based interaction priors to enforce physical plausibility and accurately model inter-human contact. Finally, the multi-view images are fused into a high-fidelity texture. Together, these components form our complete framework, HUG3D. Extensive experiments show that HUG3D significantly outperforms both single-human and existing multi-human methods, producing physically plausible, high-fidelity 3D reconstructions of interacting people from a single image. Project page: https://jongheean11.github.io/HUG3D_project