DSERT-RoLL: Robust Multi-Modal Perception for Diverse Driving Conditions with Stereo Event-RGB-Thermal Cameras, 4D Radar, and Dual-LiDAR

arXiv cs.CV / 4/7/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces DSERT-RoLL, a new driving dataset that combines stereo event/RGB/thermal cameras with 4D radar and dual LiDAR to cover a wide range of weather and lighting conditions.
  • DSERT-RoLL includes precise 2D/3D bounding boxes with track IDs and ego-vehicle odometry, aiming to support fair benchmarking across different sensor setups.
  • The dataset is intended to reduce data scarcity for emerging sensing modalities like event cameras and 4D radar, enabling systematic study of their performance and fusion behavior.
  • It provides unified 2D/3D benchmarks, baselines for both single-modality and multimodal approaches, and protocols encouraging research on different fusion strategies and sensor combinations.
  • The authors also propose a sensor-fusion framework that maps sensor-specific cues into a unified feature space and improves 3D detection robustness under varied environmental conditions.

Abstract

In this paper, we present DSERT-RoLL, a driving dataset that incorporates stereo event, RGB, and thermal cameras together with 4D radar and dual LiDAR, collected across diverse weather and illumination conditions. The dataset provides precise 2D and 3D bounding boxes with track IDs and ego vehicle odometry, enabling fair comparisons within and across sensor combinations. It is designed to alleviate data scarcity for novel sensors such as event cameras and 4D radar and to support systematic studies of their behavior. We establish unified 3D and 2D benchmarks that enable direct comparison of characteristics and strengths across sensor families and within each family. We report baselines for representative single modality and multimodal methods and provide protocols that encourage research on different fusion strategies and sensor combinations. In addition, we propose a fusion framework that integrates sensor specific cues into a unified feature space and improves 3D detection robustness under varied weather and lighting.