Learning Explicit Continuous Motion Representation for Dynamic Gaussian Splatting from Monocular Videos

arXiv cs.CV / 3/27/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a method for high-quality dynamic Gaussian Splatting from monocular videos by explicitly modeling continuous position and orientation deformations of dynamic Gaussians.
  • It represents motion using an SE(3) B-spline basis with a compact set of control points, aiming to capture complex dynamics more effectively than prior approaches.
  • An adaptive control mechanism dynamically adjusts the number of motion bases and control points to improve computational efficiency.
  • To handle long-interval motion interference, the method uses a soft segment reconstruction strategy, and it employs a multi-view diffusion model to reduce overfitting to training views.
  • Experiments report improved performance over state-of-the-art methods for novel view synthesis, with code provided on GitHub.

Abstract

We present an approach for high-quality dynamic Gaussian Splatting from monocular videos. To this end, we in this work go one step further beyond previous methods to explicitly model continuous position and orientation deformation of dynamic Gaussians, using an SE(3) B-spline motion bases with a compact set of control points. To improve computational efficiency while enhancing the ability to model complex motions, an adaptive control mechanism is devised to dynamically adjust the number of motion bases and control points. Besides, we develop a soft segment reconstruction strategy to mitigate long-interval motion interference, and employ a multi-view diffusion model to provide multi-view cues for avoiding overfitting to training views. Extensive experiments demonstrate that our method outperforms state-of-the-art methods in novel view synthesis. Our code is available at https://github.com/hhhddddddd/se3bsplinegs.