UCS: Estimating Unseen Coverage for Improved In-Context Learning

arXiv cs.LG / 4/15/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces UCS (Unseen Coverage Selection), a training-free method for improving in-context learning by selecting demonstration sets based on how well they cover latent clusters not present in the currently selected subset.
  • UCS works by inducing discrete latent clusters from model-consistent embeddings and then estimating unrevealed clusters in a candidate subset using a Smoothed Good–Turing estimator derived from its empirical frequency spectrum.
  • The authors show UCS can be combined with existing query-dependent or query-independent selection baselines via a simple regularized objective without retraining.
  • Experiments on intent-classification and reasoning benchmarks with frontier LLMs find that adding UCS to strong baselines improves ICL accuracy by about 2–6% under the same selection budget.
  • The approach also provides interpretability by yielding insights into task- and model-level latent cluster distributions, and the authors release accompanying code on GitHub.

Abstract

In-context learning (ICL) performance depends critically on which demonstrations are placed in the prompt, yet most existing selectors prioritize heuristic notions of relevance or diversity and provide limited insight into the coverage of a demonstration set. We propose Unseen Coverage Selection (UKS), a training-free, subset-level coverage prior motivated by the principle that a good demonstration set should expose the model to latent cluster unrevealed by the currently selected subset. UCS operationalizes this idea by (1) inducing discrete latent clusters from model-consistent embeddings and (2) estimating the number of unrevealed clusters within a candidate subset via a Smoothed Good--Turing estimator from its empirical frequency spectrum. Unlike previous selection methods, UCS is coverage-based and training-free, and can be seamlessly combined with both query-dependent and query-independent selection baselines via a simple regularized objective. Experiments on multiple intent-classification and reasoning benchmarks with frontier Large Language Models show that augmenting strong baselines with UCS consistently improves ICL accuracy by up to 2-6% under the same selection budget, while also yielding insights into task- and model-level latent cluster distributions. Code is available at https://github.com/Raina-Xin/UCS.