PanoAir: A Panoramic Visual-Inertial SLAM with Cross-Time Real-World UAV Dataset

arXiv cs.RO / 4/2/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces PanoAir, a panoramic visual-inertial SLAM approach intended to improve UAV pose estimation by addressing drift and failure modes caused by limited field-of-view sensors in existing VI-SLAM methods.
  • It also releases a new real-world panoramic VI dataset covering diverse flight conditions such as varying illumination, altitudes, trajectory lengths, and motion dynamics to better stress-test SLAM in practical settings.
  • PanoAir’s framework uses panoramic feature extraction and panoramic loop closure to strengthen feature constraints and maintain global consistency for more accurate and robust localization.
  • Experiments on the new dataset and public benchmarks report improved accuracy, robustness, and consistency over prior methods.
  • The work further includes an embedded-platform deployment demonstration and provides publicly available code and dataset for replication and real-world experimentation.

Abstract

Accurate pose estimation is fundamental for unmanned aerial vehicle (UAV) applications, where Visual-Inertial SLAM (VI-SLAM) provides a cost-effective solution for localization and mapping. However, existing VI-SLAM methods mainly rely on sensors with limited fields of view (FoV), which can lead to drift and even failure in complex UAV scenarios. Although panoramic cameras provide omnidirectional perception to improve robustness, panoramic VI-SLAM and corresponding real-world datasets for UAVs remain underexplored. To address this limitation, we first construct a real-world panoramic visual-inertial dataset covering diverse flight conditions, including varying illumination, altitudes, trajectory lengths, and motion dynamics. To achieve accurate and robust pose estimation under such challenging UAV scenarios, we propose a panoramic VI-SLAM framework that exploits the omnidirectional FoV via the proposed panoramic feature extraction and panoramic loop closure, enhancing feature constraints and ensuring global consistency. Extensive experiments on both the proposed dataset and public benchmarks demonstrate that our method achieves superior accuracy, robustness, and consistency compared to existing approaches. Moreover, deployment on embedded platform validates its practical applicability, achieving comparable computational efficiency to PC implementations. The source code and dataset are publicly available at https://drive.google.com/file/d/1lG1Upn6yi-N6tYpEHAt6dfR1uhzNtWbT/view