Synthetic Dataset Generation for Partially Observed Indoor Objects

arXiv cs.CV / 4/9/2026

💬 OpinionSignals & Early TrendsModels & Research

Key Points

  • The paper introduces a Unity-based virtual scanning framework that generates realistic synthetic indoor 3D scan data by simulating scanner parameters like resolution, range, and distance-dependent noise.
  • It uses ray-based scanning from configurable viewpoints to accurately model occlusion and sensor visibility, producing partial point clouds suitable for partially observed object learning.
  • The system assigns color to point clouds using panoramic images taken at the virtual scanner pose, improving the realism of the generated scans.
  • For scalability, the scanner is connected to a procedural indoor scene generator that creates diverse rooms and furniture layouts automatically.
  • The authors release the V-Scan dataset, which includes partial object point clouds, voxel-based occlusion grids, and complete ground-truth geometry for training and evaluating 3D scene reconstruction and object completion methods.

Abstract

Learning-based methods for 3D scene reconstruction and object completion require large datasets containing partial scans paired with complete ground-truth geometry. However, acquiring such datasets using real-world scanning systems is costly and time-consuming, particularly when accurate ground truth for occluded regions is required. In this work, we present a virtual scanning framework implemented in Unity for generating realistic synthetic 3D scan datasets. The proposed system simulates the behaviour of real-world scanners using configurable parameters such as scan resolution, measurement range, and distance-dependent noise. Instead of directly sampling mesh surfaces, the framework performs ray-based scanning from virtual viewpoints, enabling realistic modelling of sensor visibility and occlusion effects. In addition, panoramic images captured at the scanner location are used to assign colours to the resulting point clouds. To support scalable dataset creation, the scanner is integrated with a procedural indoor scene generation pipeline that automatically produces diverse room layouts and furniture arrangements. Using this system, we introduce the \textit{V-Scan} dataset, which contains synthetic indoor scans together with object-level partial point clouds, voxel-based occlusion grids, and complete ground-truth geometry. The resulting dataset provides valuable supervision for training and evaluating learning-based methods for scene reconstruction and object completion.