Label Shift Estimation With Incremental Prior Update

arXiv cs.LG / 4/3/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses label shift estimation in supervised learning when the training and test label distributions differ, while assuming the conditional distribution p(x|y) stays constant (no concept drift).
  • It proposes a post-hoc method that incrementally updates the prior on each new sample and recomputes the posterior to improve estimation of the test-time label distribution p_t(y).
  • Unlike prior approaches based on moment matching with a confusion matrix or maximum-likelihood/EM over new data, the method is designed to work with weaker calibration requirements.
  • The approach targets compatibility with black-box probabilistic classifiers, making it applicable without retraining the model.
  • Experiments on CIFAR-10 and MNIST indicate the method achieves consistent gains over maximum-likelihood baselines across different calibration settings and label-shift magnitudes.

Abstract

An assumption often made in supervised learning is that the training and testing sets have the same label distribution. However, in real-life scenarios, this assumption rarely holds. For example, medical diagnosis result distributions change over time and across locations; fraud detection models must adapt as patterns of fraudulent activity shift; the category distribution of social media posts changes based on trending topics and user demographics. In the task of label shift estimation, the goal is to estimate the changing label distribution p_t(y) in the testing set, assuming the likelihood p(x|y) does not change, implying no concept drift. In this paper, we propose a new approach for post-hoc label shift estimation, unlike previous methods that perform moment matching with confusion matrix estimated from a validation set or maximize the likelihood of the new data with an expectation-maximization algorithm. We aim to incrementally update the prior on each sample, adjusting each posterior for more accurate label shift estimation. The proposed method is based on intuitive assumptions on classifiers that are generally true for modern probabilistic classifiers. The proposed method relies on a weaker notion of calibration compared to other methods. As a post-hoc approach for label shift estimation, the proposed method is versatile and can be applied to any black-box probabilistic classifier. Experiments on CIFAR-10 and MNIST show that the proposed method consistently outperforms the current state-of-the-art maximum likelihood-based methods under different calibrations and varying intensities of label shift.