Cross-Modal Reinforcement Learning for Navigation with Degraded Depth Measurements

arXiv cs.RO / 3/24/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a cross-modal navigation framework that combines depth and grayscale imagery to remain robust when depth sensors are degraded by poor lighting or reflective surfaces.
  • It introduces a Cross-Modal Wasserstein Autoencoder that learns shared latent representations by enforcing cross-modal consistency, allowing depth-relevant features to be inferred from grayscale inputs.
  • The learned representations are then used with a reinforcement learning policy to enable collision-free navigation in unstructured environments.
  • Experiments in both simulation and real-world settings show the method maintains strong performance under significant depth degradation and transfers effectively to real environments.

Abstract

This paper presents a cross-modal learning framework that exploits complementary information from depth and grayscale images for robust navigation. We introduce a Cross-Modal Wasserstein Autoencoder that learns shared latent representations by enforcing cross-modal consistency, enabling the system to infer depth-relevant features from grayscale observations when depth measurements are corrupted. The learned representations are integrated with a Reinforcement Learning-based policy for collision-free navigation in unstructured environments when depth sensors experience degradation due to adverse conditions such as poor lighting or reflective surfaces. Simulation and real-world experiments demonstrate that our approach maintains robust performance under significant depth degradation and successfully transfers to real environments.