SpectralSplat: Appearance-Disentangled Feed-Forward Gaussian Splatting for Driving Scenes

arXiv cs.CV / 4/7/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • SpectralSplat addresses a key limitation of feed-forward 3D Gaussian Splatting for driving scenes by disentangling scene geometry from transient appearance factors like lighting, weather, and time of day.
  • The method predicts color via two streams—an appearance-agnostic base and an appearance-conditioned adapted stream—both produced by a shared MLP using a global appearance embedding from DINOv2 features.
  • To train effective disentanglement, it uses paired observations synthesized by a hybrid relighting pipeline that combines physics-based intrinsic decomposition with diffusion-based generative refinement.
  • It introduces an appearance-adaptable temporal history that stores appearance-agnostic features, allowing accumulated Gaussians to be re-rendered under arbitrary target appearances while maintaining temporal consistency.
  • Experiments indicate SpectralSplat retains the reconstruction quality of the underlying Gaussian Splatting backbone while enabling controllable appearance transfer and consistent relighting across driving sequences.

Abstract

Feed-forward 3D Gaussian Splatting methods have achieved impressive reconstruction quality for autonomous driving scenes, yet they entangle scene geometry with transient appearance properties such as lighting, weather, and time of day. This coupling prevents relighting, appearance transfer, and consistent rendering across multi-traversal data captured under varying environmental conditions. We present SpectralSplat, a method that disentangles appearance from geometry within a feed-forward Gaussian Splatting framework. Our key insight is to factor color prediction into an appearance-agnostic base stream and and appearance-conditioned adapted stream, both produced by a shared MLP conditioned on a global appearance embedding derived from DINOv2 features. To enforce disentanglement, we train with paired observations generated by a hybrid relighting pipeline that combines physics-based intrinsic decomposition with diffusion based generative refinement, and supervise with complementary consistency, reconstruction, cross-appearance, and base color losses. We further introduce an appearance-adaptable temporal history that stores appearance-agnostic features, enabling accumulated Gaussians to be re-rendered under arbitrary target appearances. Experiments demonstrate that SpectralSplat preserves the reconstruction quality of the underlying backbone while enabling controllable appearance transfer and temporally consistent relighting across driving sequences.

SpectralSplat: Appearance-Disentangled Feed-Forward Gaussian Splatting for Driving Scenes | AI Navigate