NRGS: Neural Regularization for Robust 3D Semantic Gaussian Splatting

arXiv cs.CV / 4/27/2026

📰 NewsModels & Research

Key Points

  • The paper introduces a neural regularization approach to improve noisy 3D semantic fields when projecting inconsistent multi-view 2D features into 3D semantic Gaussian Splatting.
  • It addresses a core issue with vision foundation model features: they often lack cross-view constraints, causing inconsistencies across views that lead to degraded downstream performance.
  • Instead of relying on preprocessing to enforce multi-view consistency or using heavier optimization to suppress noise, the method applies a variance-aware conditional MLP directly on 3D Gaussians.
  • By using the geometric and appearance attributes of the 3D Gaussians, the model corrects semantic errors in 3D space and improves accuracy across multiple datasets.

Abstract

We propose a neural regularization method that refines the noisy 3D semantic field produced by lifting multi-view inconsistent 2D features, in order to obtain an accurate and robust 3D semantic Gaussian Splatting. The 2D features extracted from vision foundation models suffer from multi-view inconsistency due to a lack of cross-view constraints. Lifting these inconsistent features directly into 3D Gaussians results in a noisy semantic field, which degrades the performance of downstream tasks. Previous methods either focus on obtaining consistent multi-view features in the preprocessing stage or aim to mitigate noise through improved optimization strategies, often at the cost of increased preprocessing time or expensive computational overhead. In contrast, we introduce a variance-aware conditional MLP that operates directly on the 3D Gaussians, leveraging their geometric and appearance attributes to correct semantic errors in 3D space. Experiments on different datasets show that our method enhances the accuracy of lifted semantics, providing an efficient and effective approach to robust 3D semantic Gaussian Splatting.