Le MuMo JEPA: Multi-Modal Self-Supervised Representation Learning with Learnable Fusion Tokens

arXiv cs.CV / 3/26/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • Le MuMo JEPA is introduced as a multi-modal self-supervised representation learning framework that learns unified embeddings from RGB images and aligned companion modalities (notably camera-aligned LiDAR depth).
  • The method extends LeJEPA by adding learnable fusion tokens that form a latent bottleneck inside a shared transformer, with an efficient “pruned fusion” strategy that drops modality-specific tokens after an initial cross-modal attention layer.
  • It applies SIGReg regularization to the joint multimodal CLS embedding to improve representation quality for downstream tasks.
  • Driving experiments on Waymo and nuScenes show Le MuMo JEPA achieves strong performance-efficiency trade-offs versus from-scratch multimodal baselines, improving CenterNet detection and dense depth while staying competitive on segmentation.
  • The framework also transfers well to the Teledyne FLIR ADAS benchmark, delivering the best results in the study, particularly after Waymo-initialized fine-tuning, with reduced compute/memory/training time.

Abstract

Self-supervised learning has emerged as a powerful paradigm for learning visual representations without manual annotations, yet most methods still operate on a single modality and therefore miss the complementary structure available from heterogeneous sensors. We present Le MuMo JEPA, a self-supervised framework that learns unified representations from RGB images and aligned companion modalities. In our driving experiments, the second modality is camera-aligned LiDAR depth; we also evaluate RGB-thermal training and transfer on the Teledyne FLIR ADAS benchmark. Our approach extends LeJEPA to the multi-modal setting by learning fusion tokens that act as a latent bottleneck between modality-specific patch stems inside a shared transformer. Our default model employs a pruned fusion strategy: after an initial cross-modal attention layer, modality-specific tokens are dropped, forcing cross-modal information into the shared fusion-token grid as an efficient latent bottleneck before Sketched Isotropic Gaussian Regularization (SIGReg) is applied to the joint multimodal CLS embedding. On Waymo, Le MuMo JEPA gives the strongest performance-efficiency trade-off on downstream patch probes among the from-scratch multimodal baselines, improving CenterNet detection and dense depth while remaining competitive on segmentation. Under from-scratch training on nuScenes, Le MuMo JEPA remains the strongest model, and it also gives the best FLIR results, especially after Waymo-initialized fine-tuning. It also retains the best overall accuracy-efficiency balance in our study at substantially lower compute, memory, and estimated training time.