Unstable Rankings in Bayesian Deep Learning Evaluation

arXiv cs.LG / 4/28/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper shows that in Bayesian deep learning, evaluation rankings can become unreliable when training data is scarce, and the rankings can be strongly dataset-dependent in ways point estimates cannot expose.
  • It finds that there is no single universal sample-size threshold at which method rankings become stable across datasets, so conclusions require dataset-specific Bayesian posterior inference.
  • The authors introduce a Bayesian hierarchical modeling approach that treats evaluation metrics as random variables across data realizations and incorporates method-specific variances.
  • They propose using a predictive Minimum Detectable Difference (MDD) curve to determine whether an observed performance gap would be detectable at a given training size.
  • Experiments across six Bayesian deep learning methods and five regression datasets indicate that uncertainty-aware evaluation is necessary, since evidence of superiority and detectability can diverge even at the same training size.

Abstract

Standard evaluations of Bayesian deep learning methods assume that metric estimates are reliable, but we show this assumption fails under data scarcity. Method rankings are not only unreliable at small n, but also dataset-dependent in ways that point estimates cannot reveal: the same method comparison yields P(\mathrm{MCD} \prec \mathrm{Ensemble}) = 1.000 at n = 50 on one dataset and remains below 0.95 even at n = 500 on another. Across the datasets we consider, no universal sample size threshold exists, which is precisely why dataset-specific posterior inference is necessary. To address this, we use a Bayesian hierarchical model with method-specific variances to treat evaluation metrics as random variables across data realizations, and we use a predictive Minimum Detectable Difference curve to assess whether an observed gap would be detectable at a given training size. Across six Bayesian deep learning methods and five regression datasets, our results show that uncertainty-aware evaluation is necessary in low-data settings, because current evidence for method superiority and predictive detectability at the same training size can diverge substantially. Our framework provides practitioners with principled tools to determine whether their evaluation data is sufficient before drawing conclusions about method superiority.