Love Me, Love My Label: Rethinking the Role of Labels in Prompt Retrieval for Visual In-Context Learning
arXiv cs.CV / 4/7/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- Visual in-context learning (VICL) performance depends heavily on selecting the right demonstrative prompts, and existing prompt retrieval methods often ignore whether prompt labels match the query labels.
- The study finds that visually similar but label-inconsistent prompts can degrade VICL results, while stronger label consistency between query and prompts correlates with better outcomes.
- To address this, the authors propose LaPR (Label-aware Prompt Retrieval), which builds an image-label joint representation to incorporate label cues explicitly during prompt selection.
- LaPR also introduces a mixture-of-experts mechanism with query-adaptive routing to handle missing query labels at test time, training experts and a router using both VICL performance-guided and label-guided contrastive losses.
- Experiments across in-context segmentation, detection, and colorization show consistent improvements over prior approaches, with good generalization across feature extractors and cross-fold settings; code is publicly available.
Related Articles

Black Hat Asia
AI Business
[R] The ECIH: Model Modeling Agentic Identity as an Emergent Relational State [R]
Reddit r/MachineLearning
Google DeepMind Unveils Project Genie: The Dawn of Infinite AI-Generated Game Worlds
Dev.to
Artificial Intelligence and Life in 2030: The One Hundred Year Study onArtificial Intelligence
Dev.to
Stop waiting for Java to rebuild! AI IDEs + Zero-Latency Hot Reload = Magic
Dev.to