Multimodal Learning on Low-Quality Data with Conformal Predictive Self-Calibration

arXiv cs.CV / 5/6/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies multimodal learning under low-quality data, attributing both modality imbalance and noisy corruption to miscalibrated predictive uncertainty about the reliability of modalities and instances during training.
  • It introduces Conformal Predictive Self-Calibration (CPSC), a unified framework that uses conformal prediction to enable self-guided, online calibration while the model trains.
  • CPSC combines Representation Self-Calibration (decomposing unimodal features and selectively fusing the most reliable parts) with Gradient Self-Calibration (reweighting or redirecting gradient flow using instance-wise reliability scores).
  • The approach includes a self-update mechanism for the conformal predictor so the calibration components co-evolve consistently throughout optimization.
  • Experiments on six benchmark datasets in both imbalanced and noisy regimes show CPSC outperforms prior state-of-the-art multimodal methods, and the authors release code on GitHub.

Abstract

Multimodal learning often grapples with the challenge of low-quality data, which predominantly manifests as two facets: modality imbalance and noisy corruption. While these issues are often studied in isolation, we argue that they share a common root in the predictive uncertainty towards the reliability of individual modalities and instances during learning. In this paper, we propose a unified framework, termed Conformal Predictive Self-Calibration (CPSC), which leverages conformal prediction to equip the model with the ability to perform self-guided calibration on-the-fly. The core of our proposed CPSC lies in a novel self-calibrating training loop that seamlessly integrates two key modules: (1) Representation Self-Calibration, which decomposes unimodal features into components, and selectively fuses the most robust ones identified by a conformal predictor to enhance feature resilience. (2) Gradient Self-Calibration, which recalibrates the gradient flow during backpropagation based on instance-wise reliability scores, steering the optimization towards more trustworthy directions. Furthermore, we also devise a self-update strategy for the conformal predictor to ensure the entire system co-evolves consistently throughout the training process. Extensive experiments on six benchmark datasets under both imbalanced and noisy settings demonstrate that our CPSC framework consistently outperforms existing state-of-the-art methods. Our code is available at https://github.com/XunCHN/CPSC.