Robust Visual SLAM for UAV Navigation in GPS-Denied and Degraded Environments: A Multi-Paradigm Evaluation and Deployment Study

arXiv cs.RO / 5/6/2026

📰 NewsDeveloper Stack & InfrastructureSignals & Early TrendsModels & Research

Key Points

  • The study provides a systematic comparison of five Visual SLAM systems (ORB-SLAM3, DPVO, DROID-SLAM, DUSt3R, and MASt3R) across multiple modeling paradigms under GPS-denied, visually degraded conditions.
  • Across benchmark datasets and a custom monocular indoor dataset with controlled degradations (low light, dust haze, motion blur, and combinations), learning-based approaches show much more robust localization than the classical method, with ORB-SLAM3 failing severely (62.4% overall TSR and 0% under dense haze).
  • MASt3R achieves the lowest degraded absolute trajectory error (0.027 m), while DUSt3R attains the highest tracking success rate (96.5%), indicating different strengths depending on the evaluation metric.
  • DPVO delivers the best efficiency–robustness trade-off for embedded UAV use, achieving 18.6 FPS, 3.1 GB GPU memory usage, and 86.1% TSR.
  • The paper also evaluates embedded deployment on NVIDIA Jetson platforms and provides practical selection guidelines for choosing SLAM systems under SWaP (size, weight, power) constraints in UAV operations.

Abstract

Reliable localization in GPS-denied, visually degraded environments is critical for autonomous UAV opera- tions. This paper presents a systematic comparative evaluation of five V-SLAM systems ORB-SLAM3, DPVO, DROID-SLAM, DUSt3R, and MASt3R spanning classical, deep learning, recurrent, and Vision Transformer (ViT) paradigms. Experiments are conducted on curated sequences from four public benchmarks (TUM RGB-D, EuRoC MAV, UMA-VI, SubT-MRS) and a custom monocular indoor dataset under five controlled degradation conditions (normal, low light, dust haze, motion blur, and combined), with sub-millimeter Vicon ground truth. Results show that ORB-SLAM3 fails critically under severe degradation (62.4% overall TSR; 0% under dense haze), while learning-based methods remain robust: MASt3R achieves the lowest degraded ATE (0.027 m) and DUSt3R the highest tracking success (96.5%). DPVO offers the best efficiency robustness trade-off (18.6 FPS, 3.1 GB GPU memory, 86.1% TSR), making it the preferred choice for memory-constrained embedded platforms. Embedded deployment analysis across NVIDIA Jetson platforms provides actionable guidelines for SLAM selection under SWaP-constrained UAV scenarios.