PTC-Depth: Pose-Refined Monocular Depth Estimation with Temporal Consistency
arXiv cs.CV / 4/3/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper addresses a key limitation in monocular depth estimation: lack of temporal consistency across consecutive video frames, which can cause depth jitter and failures during abrupt depth-range changes.
- It proposes PTC-Depth, a consistency-aware framework that uses wheel odometry to stabilize camera pose and employs optical flow to compute triangulation-based sparse depth.
- Sparse depth updates a recursive Bayesian estimate of metric scale over time, which is then used to rescale the relative depth output from a pre-trained depth “foundation” model.
- Experiments on KITTI, TartanAir, MS2, and an additional dataset show improved robustness and accuracy for metric depth estimation under temporal dynamics.
Related Articles

Why I built an AI assistant that doesn't know who you are
Dev.to

DenseNet Paper Walkthrough: All Connected
Towards Data Science

Meta Adaptive Ranking Model: What Instagram Advertisers Gain in 2026 | MKDM
Dev.to

The Facebook insider building content moderation for the AI era
TechCrunch
Qwen3.5 vs Gemma 4: Benchmarks vs real world use?
Reddit r/LocalLLaMA