REALITrees: Rashomon Ensemble Active Learning for Interpretable Trees

arXiv stat.ML / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • REALITrees introduces Rashomon Ensembled Active Learning (REAL), a method that improves active learning by building a committee from the Rashomon Set of all near-optimal sparse decision-tree models rather than relying on perturbation-based disagreement as in Query-by-Committee (QBC).
  • The approach reduces functional redundancy in the committee using a PAC-Bayesian framework with a Gibbs posterior that weights committee members by empirical risk, aiming to better characterize the plausible hypothesis space.
  • For sparse decision trees, the method exactly enumerates the Rashomon Set via recent algorithmic advances, enabling a more faithful committee than randomized ensembles.
  • Experiments on synthetic data and established active learning baselines show REAL outperforming randomized ensembles, with the largest gains in moderately noisy settings through faster convergence enabled by expanded model multiplicity.

Abstract

Active learning reduces labeling costs by selecting samples that maximize information gain. A dominant framework, Query-by-Committee (QBC), typically relies on perturbation-based diversity by inducing model disagreement through random feature subsetting or data blinding. While this approximates one notion of epistemic uncertainty, it sacrifices direct characterization of the plausible hypothesis space. We propose the complementary approach: Rashomon Ensembled Active Learning (REAL) which constructs a committee by exhaustively enumerating the Rashomon Set of all near-optimal models. To address functional redundancy within this set, we adopt a PAC-Bayesian framework using a Gibbs posterior to weight committee members by their empirical risk. Leveraging recent algorithmic advances, we exactly enumerate this set for the class of sparse decision trees. Across synthetic and established active learning baselines, REAL outperforms randomized ensembles, particularly in moderately noisy environments where it strategically leverages expanded model multiplicity to achieve faster convergence.