U-ViLAR: Uncertainty-Aware Visual Localization for Autonomous Driving via Differentiable Association and Registration

arXiv cs.RO / 4/28/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • U-ViLAR is a new uncertainty-aware visual localization framework for autonomous driving in urban environments where GNSS signals are degraded or unreliable.
  • The approach maps both image features and map information into Bird’s-Eye-View (BEV) space to improve spatial consistency with HD or navigation maps.
  • It introduces Perceptual Uncertainty-guided Association to reduce localization errors caused by uncertainty in perception.
  • It also proposes Localization Uncertainty-guided Registration to limit errors introduced by uncertainty during the localization refinement stage.
  • Experiments report state-of-the-art results across multiple localization tasks, with additional validation on large-scale autonomous driving fleets showing stable performance in challenging scenarios.

Abstract

Accurate localization using visual information is a critical yet challenging task, especially in urban environments where nearby buildings and construction sites significantly degrade GNSS (Global Navigation Satellite System) signal quality. This issue underscores the importance of visual localization techniques in scenarios where GNSS signals are unreliable. This paper proposes U-ViLAR, a novel uncertainty-aware visual localization framework designed to address these challenges while enabling adaptive localization using high-definition (HD) maps or navigation maps. Specifically, our method first extracts features from the input visual data and maps them into Bird's-Eye-View (BEV) space to enhance spatial consistency with the map input. Subsequently, we introduce: a) Perceptual Uncertainty-guided Association, which mitigates errors caused by perception uncertainty, and b) Localization Uncertainty-guided Registration, which reduces errors introduced by localization uncertainty. By effectively balancing the coarse-grained large-scale localization capability of association with the fine-grained precise localization capability of registration, our approach achieves robust and accurate localization. Experimental results demonstrate that our method achieves state-of-the-art performance across multiple localization tasks. Furthermore, our model has undergone rigorous testing on large-scale autonomous driving fleets and has demonstrated stable performance in various challenging urban scenarios.