Active In-Context Learning for Tabular Foundation Models

arXiv cs.LG / 3/31/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that traditional active learning struggles in tabular cold-start settings because uncertainty estimates become unreliable with very few labeled samples.
  • It proposes Tabular Active In-Context Learning (Tab-AICL), leveraging tabular foundation models (e.g., TabPFN) whose in-context learning can optimize the labeled context without updating model weights.
  • The authors formalize four acquisition strategies for selecting new labels: uncertainty (TabPFN-Margin), diversity (TabPFN-Coreset), an uncertainty-diversity hybrid (TabPFN-Hybrid), and a scalable two-stage shortlist-then-select approach (TabPFN-Proxy-Hybrid).
  • Experiments on 20 classification benchmarks show Tab-AICL improves cold-start sample efficiency versus retrained gradient-boosting baselines (CatBoost/XGBoost margin), achieving gains measured by normalized AULC up to 100 labeled samples.
  • The work positions tabular foundation model calibration plus context-optimized acquisition as a promising route to reduce labeling costs in practical, low-data regimes.

Abstract

Active learning (AL) reduces labeling cost by querying informative samples, but in tabular settings its cold-start gains are often limited because uncertainty estimates are unreliable when models are trained on very few labels. Tabular foundation models such as TabPFN provide calibrated probabilistic predictions via in-context learning (ICL), i.e., without task-specific weight updates, enabling an AL regime in which the labeled context - rather than parameters - is iteratively optimized. We formalize Tabular Active In-Context Learning (Tab-AICL) and instantiate it with four acquisition rules: uncertainty (TabPFN-Margin), diversity (TabPFN-Coreset), an uncertainty-diversity hybrid (TabPFN-Hybrid), and a scalable two-stage method (TabPFN-Proxy-Hybrid) that shortlists candidates using a lightweight linear proxy before TabPFN-based selection. Across 20 classification benchmarks, Tab-AICL improves cold-start sample efficiency over retrained gradient-boosting baselines (CatBoost-Margin and XGBoost-Margin), measured by normalized AULC up to 100 labeled samples.