ROS 2-Based LiDAR Perception Framework for Mobile Robots in Dynamic Production Environments, Utilizing Synthetic Data Generation, Transformation-Equivariant 3D Detection and Multi-Object Tracking

arXiv cs.RO / 4/3/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a ROS 2-based LiDAR perception framework for mobile robots that targets 6D pose estimation and multi-object tracking in dynamic industrial production environments.
  • It trains a Transformation-Equivariant 3D detector using synthetic data to reduce dependency on real-world data while improving noise robustness and spatiotemporal consistency.
  • The framework integrates multi-object tracking using “center poses,” improving detection-to-tracking continuity over standalone pose estimation.
  • On 72 motion-capture-evaluated scenarios, the authors report IoU of 62.6% for standalone 6D pose estimation and 83.12% after adding multi-object tracking.
  • The system also achieves 91.12% Higher Order Tracking Accuracy, indicating stronger robustness and versatility for LiDAR-based perception in industrial mobile manipulators.

Abstract

Adaptive robots in dynamic production environments require robust perception capabilities, including 6D pose estimation and multi-object tracking. To address limitations in real-world data dependency, noise robustness, and spatiotemporal consistency, a LiDAR framework based on the Robot Operating System integrating a synthetic-data-trained Transformation-Equivariant 3D Detection with multi-object-tracking leveraging center poses is proposed. Validated across 72 scenarios with motion capture technology, overall results yield an Intersection over Union of 62.6% for standalone pose estimation, rising to 83.12% with multi-object-tracking integration. Our LiDAR-based framework achieves 91.12% of Higher Order Tracking Accuracy, advancing robustness and versatility of LiDAR-based perception systems for industrial mobile manipulators.