Holistic Optimal Label Selection for Robust Prompt Learning under Partial Labels

arXiv cs.CV / 4/9/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Holistic Optimal Label Selection (HopS) to improve prompt learning for vision-language models when training data has only partial/ambiguous labels.
  • HopS uses a local density-based filtering strategy over nearest-neighbor label candidates, combining label frequency and softmax confidence to pick plausible labels.
  • It also adds a global label assignment objective using optimal transport to align a uniform sampling distribution with candidate label distributions across a batch by minimizing expected transport cost.
  • Experiments on eight benchmark datasets show HopS consistently boosts performance under partial supervision and surpasses prior baselines, indicating stronger robustness in weakly supervised settings.

Abstract

Prompt learning has gained significant attention as a parameter-efficient approach for adapting large pre-trained vision-language models to downstream tasks. However, when only partial labels are available, its performance is often limited by label ambiguity and insufficient supervisory information. To address this issue, we propose Holistic Optimal Label Selection (HopS), leveraging the generalization ability of pre-trained feature encoders through two complementary strategies. First, we design a local density-based filter that selects the top frequent labels from the nearest neighbors' candidate sets and uses the softmax scores to identify the most plausible label, capturing structural regularities in the feature space. Second, we introduce a global selection objective based on optimal transport that maps the uniform sampling distribution to the candidate label distributions across a batch. By minimizing the expected transport cost, it can determine the most likely label assignments. These two strategies work together to provide robust label selection from both local and global perspectives. Extensive experiments on eight benchmark datasets show that HopS consistently improves performance under partial supervision and outperforms all baselines. Those results highlight the merit of holistic label selection and offer a practical solution for prompt learning in weakly supervised settings.