AI Navigate

Learning Geometric and Photometric Features from Panoramic LiDAR Scans for Outdoor Place Categorization

arXiv cs.CV / 3/16/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes CNN-based methods that use panoramic LiDAR depth and reflectance images as inputs for outdoor place categorization.
  • It introduces the Multi-modal Panoramic 3D Outdoor (MPO) dataset, collected with two LiDARs and labeled into six outdoor place categories (e.g., coast, forest, parking areas, residential, urban).
  • The approach demonstrates that fusing depth and reflectance modalities yields improvements over traditional methods for outdoor place recognition.
  • The authors analyze the learned network features through visualizations to understand what geometric and photometric cues the model uses.
  • This work targets autonomous robots and vehicles, addressing challenges like long-range cues, illumination changes, and occlusions in outdoor environments.

Abstract

Semantic place categorization, which is one of the essential tasks for autonomous robots and vehicles, allows them to have capabilities of self-decision and navigation in unfamiliar environments. In particular, outdoor places are more difficult targets than indoor ones due to perceptual variations, such as dynamic illuminance over twenty-four hours and occlusions by cars and pedestrians. This paper presents a novel method of categorizing outdoor places using convolutional neural networks (CNNs), which take omnidirectional depth/reflectance images obtained by 3D LiDARs as the inputs. First, we construct a large-scale outdoor place dataset named Multi-modal Panoramic 3D Outdoor (MPO) comprising two types of point clouds captured by two different LiDARs. They are labeled with six outdoor place categories: coast, forest, indoor/outdoor parking, residential area, and urban area. Second, we provide CNNs for LiDAR-based outdoor place categorization and evaluate our approach with the MPO dataset. Our results on the MPO dataset outperform traditional approaches and show the effectiveness in which we use both depth and reflectance modalities. To analyze our trained deep networks we visualize the learned features.