Uncertainty Estimation for the Open-Set Text Classification systems

arXiv cs.AI / 4/13/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies open-set text classification (OSTC), where inputs must be assigned to known classes or rejected as unknown, emphasizing the need for accurate uncertainty estimation.
  • It adapts the Holistic Uncertainty Estimation (HolUE) approach to the text domain by modeling two distinct error sources: text uncertainty from poorly formulated queries and gallery uncertainty from ambiguous data distributions.
  • By capturing these uncertainty components, the method aims to predict when the classifier is likely to make recognition errors and trigger rejection accordingly.
  • The authors introduce a new OSTC benchmark and run extensive experiments across authorship attribution, intent, and topic classification datasets.
  • Experimental results show HolUE improves Prediction Rejection Ratio (PRR) substantially over a quality-based SCF baseline (e.g., up to 365% on Yahoo Answers), and the code/protocols are released publicly.

Abstract

Accurate uncertainty estimation is essential for building robust and trustworthy recognition systems. In this paper, we consider the open-set text classification (OSTC) task - and uncertainty estimation for it. For OSTC a text sample should be classified as one of the existing classes or rejected as unknown. To account for the different uncertainty types encountered in OSTC, we adapt the Holistic Uncertainty Estimation (HolUE) method for the text domain. Our approach addresses two major causes of prediction errors in text recognition systems: text uncertainty that stems from ill formulated queries and gallery uncertainty that is related the ambiguity of data distribution. By capturing these sources, it becomes possible to predict when the system will make a recognition error. We propose a new OSTC benchmark and conduct extensive experiments on a wide range of data, utilizing the authorship attribution, intent and topic classification datasets. HolUE achieves 40-365% improvement in Prediction Rejection Ratio (PRR) over the quality-based SCF baseline across datasets: 365% on Yahoo Answers (0.79 vs 0.17 at FPIR 0.1), 347% on DBPedia (0.85 vs 0.19), 240% on PAN authorship attribution (0.51 vs 0.15 at FPIR 0.5), and 40% on CLINC150 intent classification (0.73 vs~0.52). We make public our code and protocols https://github.com/Leonid-Erlygin/text_uncertainty.git