Flow4DGS-SLAM: Optical Flow-Guided 4D Gaussian Splatting SLAM

arXiv cs.CV / 4/27/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Flow4DGS-SLAM, a framework for SLAM in dynamic environments that uses optical flow to guide 4D Gaussian splatting reconstructions.
  • It generates category-agnostic motion masks by fitting a camera ego-motion model to decompose optical flow, separating dynamic from static Gaussians and providing flow-guided camera pose initialization.
  • To improve training efficiency for dynamic 3D Gaussian splatting, it models temporal centers at keyframes, propagates them using 3D scene flow priors, and uses an adaptive insertion strategy for dynamic initialization.
  • It characterizes temporal opacity and rotation with a Gaussian Mixture Model (GMM) to better learn complex motion dynamics, and reports state-of-the-art results across tracking, dynamic reconstruction, and training speed.

Abstract

Handling the dynamic environments is a significant research challenge in Visual Simultaneous Localization and Mapping (SLAM). Recent research combines 3D Gaussian Splatting (3DGS) with SLAM to achieve both robust camera pose estimation and photorealistic renderings. However, using SLAM to efficiently reconstruct both static and dynamic regions remains challenging. In this work, we propose an efficient framework for dynamic 3DGS SLAM guided by optical flow. Using the input depth and prior optical flow, we first propose a category-agnostic motion mask generation strategy by fitting a camera ego-motion model to decompose the optical flow. This module separates dynamic and static Gaussians and simultaneously provides flow-guided camera pose initialization. We boost the training speed of dynamic 3DGS by explicitly modeling their temporal centers at keyframes. These centers are propagated using 3D scene flow priors and are dynamically initialized with an adaptive insertion strategy. Alongside this, we model the temporal opacity and rotation using a Gaussian Mixture Model (GMM) to adaptively learn the complex dynamics. The empirical results demonstrate our state-of-the-art performance in tracking, dynamic reconstruction, and training efficiency.