Sparse-View 3D Gaussian Splatting in the Wild

arXiv cs.CV / 5/1/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces “Sparse-View 3D Gaussian Splatting in the Wild,” a sparse-view 3D synthesis framework designed for unconstrained real-world scenes that include distractors.
  • It proposes reference-guided view refinement that uses a diffusion model together with a transient mask and a reference image to improve 3D representation quality and reduce rendering artifacts.
  • To better handle sparsity in the Gaussian field, the method adds pseudo-view generation and a sparsity-aware Gaussian replication strategy to amplify Gaussians in sparse regions.
  • Experiments on public datasets show consistent improvements over existing approaches, reporting notable gains in PSNR (+17.2%), SSIM (+10.8%), and LPIPS (-4.0%), while producing high-fidelity 3D renderings.
  • The authors position the approach as enabling high-quality 3D reconstruction in the wild without labor-intensive data acquisition, and provide a project page for implementation resources.

Abstract

We propose a 3D novel sparse-view synthesis framework for unconstrained real-world scenarios that contain distractors. Unlike existing methods that primarily perform novel-view synthesis from a sparse set of constrained images without transient elements or leverage unconstrained dense image collections to enhance 3D representation in real-world scenarios, our method not only effectively tackles sparse unconstrained image collections, but also shows high-quality 3D rendering results. To do this, we introduce reference-guided view refinement with a diffusion model using a transient mask and a reference image to enhance the 3D representation and mitigate artifacts in rendered views. Furthermore, we address sparse regions in the Gaussian field via pseudo-view generation along with a sparsity-aware Gaussian replication strategy to amplify Gaussians in the sparse regions. Extensive experiments on publicly available datasets demonstrate that our methodology consistently outperforms existing methods (e.g., PSNR - 17.2%, SSIM - 10.8%, LPIPS - 4.0%) and provides high-fidelity 3D rendering results. This advancement paves the way for realizing unconstrained real-world scenarios without labor-intensive data acquisition. Our project page is available at \href{https://robotic-vision-lab.github.io/SaveWildGS/}{here}