UniDAC: Universal Metric Depth Estimation for Any Camera

arXiv cs.CV / 3/31/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • UniDAC is a new monocular metric depth estimation framework designed to generalize across diverse camera types (including fisheye and 360°) using a single model rather than multiple domain-specific models.
  • The method improves cross-camera robustness by decoupling metric depth estimation into relative depth prediction and spatially varying scale estimation.
  • It introduces a lightweight Depth-Guided Scale Estimation module that upsamples a coarse scale map to higher resolution using the relative depth map to handle local scale variations.
  • UniDAC also proposes a distortion-aware positional embedding (RoPE-φ) that accounts for spatial warping in equi-rectangular projections via latitude-aware weighting.
  • The paper reports state-of-the-art performance for cross-camera generalization, consistently outperforming prior approaches across the evaluated datasets.

Abstract

Monocular metric depth estimation (MMDE) is a core challenge in computer vision, playing a pivotal role in real-world applications that demand accurate spatial understanding. Although prior works have shown promising zero-shot performance in MMDE, they often struggle with generalization across diverse camera types, such as fisheye and 360^\circ cameras. Recent advances have addressed this through unified camera representations or canonical representation spaces, but they require either including large-FoV camera data during training or separately trained models for different domains. We propose UniDAC, an MMDE framework that presents universal robustness in all domains and generalizes across diverse cameras using a single model. We achieve this by decoupling metric depth estimation into relative depth prediction and spatially varying scale estimation, enabling robust performance across different domains. We propose a lightweight Depth-Guided Scale Estimation module that upsamples a coarse scale map to high resolution using the relative depth map as guidance to account for local scale variations. Furthermore, we introduce RoPE-\phi, a distortion-aware positional embedding that respects the spatial warping in Equi-Rectangular Projections (ERP) via latitude-aware weighting. UniDAC achieves state of the art (SoTA) in cross-camera generalization by consistently outperforming prior methods across all datasets.