Real-Time Structural Detection for Indoor Navigation from 3D LiDAR Using Bird's-Eye-View Images

arXiv cs.RO / 3/23/2026

📰 NewsTools & Practical UsageModels & Research

Key Points

  • The paper proposes a lightweight, real-time perception pipeline that projects 3D LiDAR data into 2D Bird's-Eye-View images to enable efficient structural detection for indoor navigation on resource-constrained robots.
  • It systematically evaluates feature extraction strategies, including classical geometric methods (Hough Transform, RANSAC, and LSD) and a deep learning detector based on YOLO-OBB, highlighting trade-offs in robustness and speed.
  • The YOLO-OBB detector delivers the best balance of robustness and computational efficiency, achieving end-to-end 10 Hz operation on a low-power SBC without GPU acceleration and filtering cluttered observations.
  • A spatiotemporal fusion module integrates detections across frames to improve stability, with experiments on a standard mobile robotic platform demonstrating real-time performance and method limitations.

Abstract

Efficient structural perception is essential for mapping and autonomous navigation on resource-constrained robots. Existing 3D methods are computationally prohibitive, while traditional 2D geometric approaches lack robustness. This paper presents a lightweight, real-time framework that projects 3D LiDAR data into 2D Bird's-Eye-View (BEV) images to enable efficient detection of structural elements relevant to mapping and navigation. Within this representation, we systematically evaluate several feature extraction strategies, including classical geometric techniques (Hough Transform, RANSAC, and LSD) and a deep learning detector based on YOLO-OBB. The resulting detections are integrated through a spatiotemporal fusion module that improves stability and robustness across consecutive frames. Experiments conducted on a standard mobile robotic platform highlight clear performance trade-offs. Classical methods such as Hough and LSD provide fast responses but exhibit strong sensitivity to noise, with LSD producing excessive segment fragmentation that leads to system congestion. RANSAC offers improved robustness but fails to meet real-time constraints. In contrast, the YOLO-OBB-based approach achieves the best balance between robustness and computational efficiency, maintaining an end-to-end latency (satisfying 10 Hz operation) while effectively filtering cluttered observations in a low-power single-board computer (SBC) without using GPU acceleration. The main contribution of this work is a computationally efficient BEV-based perception pipeline enabling reliable real-time structural detection from 3D LiDAR on resource-constrained robotic platforms that cannot rely on GPU-intensive processing.