ProbeMDE: Uncertainty-Guided Active Proprioception for Monocular Depth Estimation in Surgical Robotics

arXiv cs.RO / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • ProbeMDEは、手術シーンのような不確実性が高い環境でのモノキュラー深度推定(MDE)を、RGB画像に加えて疎なプロプリオセプティブ(触覚)計測を組み合わせて改善する費用意識型のアクティブセンシング手法として提案されます。
  • アンサンブルのMDEモデルにより密な深度地図を生成し、予測の不確実性はアンサンブル分散で定量化、さらに不確実性の勾配を候補計測位置に対して評価します。
  • 触れて得る計測位置の選択でモード崩壊を防ぐために、候補位置の不確実性勾配マップ上でStein Variational Gradient Descent(SVGD)を用いて最大限情報量の多い地点を探索します。
  • シミュレーションと実機実験(中心気道閉塞の外科用ファントム)で検証し、従来手法より標準の深度推定指標で精度が高く、必要なプロプリオセプティブ計測数も抑えられることを示しています。

Abstract

Monocular depth estimation (MDE) provides a useful tool for robotic perception, but its predictions are often uncertain and inaccurate in challenging environments such as surgical scenes where textureless surfaces, specular reflections, and occlusions are common. To address this, we propose ProbeMDE, a cost-aware active sensing framework that combines RGB images with sparse proprioceptive measurements for MDE. Our approach utilizes an ensemble of MDE models to predict dense depth maps conditioned on both RGB images and on a sparse set of known depth measurements obtained via proprioception, where the robot has touched the environment in a known configuration. We quantify predictive uncertainty via the ensemble's variance and measure the gradient of the uncertainty with respect to candidate measurement locations. To prevent mode collapse while selecting maximally informative locations to propriocept (touch), we leverage Stein Variational Gradient Descent (SVGD) over this gradient map. We validate our method in both simulated and physical experiments on central airway obstruction surgical phantoms. Our results demonstrate that our approach outperforms baseline methods across standard depth estimation metrics, achieving higher accuracy while minimizing the number of required proprioceptive measurements. Project page: https://brittonjordan.github.io/probe_mde/