RayMap3R: Inference-Time RayMap for Dynamic 3D Reconstruction

arXiv cs.CV / 3/24/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces RayMap3R, a training-free streaming framework for dynamic 3D reconstruction that jointly estimates scene geometry and camera poses from RGB video in real time.
  • It addresses artifacts and drift caused by moving objects by using an inference-time dual-branch scheme that detects dynamic regions by contrasting RayMap-based and image-based predictions.
  • RayMap3R suppresses dynamic interference during memory updates to improve stability of reconstruction over time.
  • The method adds reset metric alignment and state-aware smoothing to maintain metric consistency and stabilize predicted camera trajectories.
  • Experiments report state-of-the-art performance versus existing streaming approaches on multiple dynamic-scene benchmarks.

Abstract

Streaming feed-forward 3D reconstruction enables real-time joint estimation of scene geometry and camera poses from RGB images. However, without explicit dynamic reasoning, streaming models can be affected by moving objects, causing artifacts and drift. In this work, we propose RayMap3R, a training-free streaming framework for dynamic scene reconstruction. We observe that RayMap-based predictions exhibit a static-scene bias, providing an internal cue for dynamic identification. Based on this observation, we construct a dual-branch inference scheme that identifies dynamic regions by contrasting RayMap and image predictions, suppressing their interference during memory updates. We further introduce reset metric alignment and state-aware smoothing to preserve metric consistency and stabilize predicted trajectories. Our method achieves state-of-the-art performance among streaming approaches on dynamic scene reconstruction across multiple benchmarks.