Single-Eye View: Monocular Real-time Perception Package for Autonomous Driving

arXiv cs.CV / 3/24/2026

📰 NewsSignals & Early TrendsModels & Research

Key Points

  • The paper proposes LRHPerception, a real-time monocular (single-camera) perception package for autonomous driving designed to improve computational efficiency without sacrificing scene understanding quality.
  • It unifies end-to-end learning efficiency with ideas from local mapping by producing a five-channel tensor that includes RGB, road segmentation, and pixel-level depth, alongside object detection and trajectory prediction.
  • Reported results show improved performance across object tracking/prediction, road segmentation, and depth estimation while running at 29 FPS on a single GPU.
  • The authors claim a 555% speedup versus the fastest mapping-based approach, indicating a substantial reduction in runtime cost for monocular perception pipelines.

Abstract

Amidst the rapid advancement of camera-based autonomous driving technology, effectiveness is often prioritized with limited attention to computational efficiency. To address this issue, this paper introduces LRHPerception, a real-time monocular perception package for autonomous driving that uses single-view camera video to interpret the surrounding environment. The proposed system combines the computational efficiency of end-to-end learning with the rich representational detail of local mapping methodologies. With significant improvements in object tracking and prediction, road segmentation, and depth estimation integrated into a unified framework, LRHPerception processes monocular image data into a five-channel tensor consisting of RGB, road segmentation, and pixel-level depth estimation, augmented with object detection and trajectory prediction. Experimental results demonstrate strong performance, achieving real-time processing at 29 FPS on a single GPU, representing a 555% speedup over the fastest mapping-based approach.