Contrastive Conformal Sets

arXiv cs.LG / 3/30/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper extends conformal prediction to contrastive learning by building minimum-volume covering sets in the learned semantic embedding space with learnable generalized multi-norm constraints.
  • It targets user-specified coverage of positive samples (distribution-free) while also improving the ability to exclude negative samples, framing negative exclusion via the geometry and volume of the covering sets.
  • The authors provide theoretical support that volume minimization can act as a proxy for negative exclusion, allowing the method to work even when negative pairs are unavailable.
  • Experiments on both simulated and real image datasets show better inclusion–exclusion trade-offs than standard distance-based conformal baselines.

Abstract

Contrastive learning produces coherent semantic feature embeddings by encouraging positive samples to cluster closely while separating negative samples. However, existing contrastive learning methods lack principled guarantees on coverage within the semantic feature space. We extend conformal prediction to this setting by introducing minimum-volume covering sets equipped with learnable generalized multi-norm constraints. We propose a method that constructs conformal sets guaranteeing user-specified coverage of positive samples while maximizing negative sample exclusion. We establish theoretically that volume minimization serves as a proxy for negative exclusion, enabling our approach to operate effectively even when negative pairs are unavailable. The positive inclusion guarantee inherits the distribution-free coverage property of conformal prediction, while negative exclusion is maximized through learned set geometry optimized on a held-out training split. Experiments on simulated and real-world image datasets demonstrate improved inclusion-exclusion trade-offs compared to standard distance-based conformal baselines.