AI Navigate

BEV-SLD: Self-Supervised Scene Landmark Detection for Global Localization with LiDAR Bird's-Eye View Images

arXiv cs.CV / 3/19/2026

📰 NewsModels & Research

Key Points

  • BEV-SLD introduces a self-supervised LiDAR global localization method that uses bird's-eye-view images to discover scene-specific landmarks at a prescribed spatial density.
  • It uses a consistency loss to align learnable global landmark coordinates with per-frame heatmaps, yielding stable landmark detections across the scene.
  • The method achieves robust localization across campus, industrial, and forest environments and shows strong performance versus state-of-the-art methods.
  • By focusing on scene-specific landmarks rather than scene-agnostic cues, BEV-SLD aims to improve robustness and accuracy for LiDAR-based localization in varied environments.

Abstract

We present BEV-SLD, a LiDAR global localization method building on the Scene Landmark Detection (SLD) concept. Unlike scene-agnostic pipelines, our self-supervised approach leverages bird's-eye-view (BEV) images to discover scene-specific patterns at a prescribed spatial density and treat them as landmarks. A consistency loss aligns learnable global landmark coordinates with per-frame heatmaps, yielding consistent landmark detections across the scene. Across campus, industrial, and forest environments, BEV-SLD delivers robust localization and achieves strong performance compared to state-of-the-art methods.