Epistemic Errors of Imperfect Multitask Learners When Distributions Shift

arXiv stat.ML / 4/1/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies uncertainty-aware learning methods (e.g., Bayesian neural networks) that output uncertainty estimates rather than only point predictions, focusing specifically on errors caused by epistemic (reducible) uncertainty.
  • It introduces a principled definition of epistemic error and derives a decomposition-based epistemic error bound applicable to imperfect multitask learning when data distributions shift between source tasks and target tasks.
  • The bound attributes epistemic errors to multiple factors, including which source tasks generate training data, how target data differs from those tasks under shift, and whether the learner can accurately characterize the source-data process.
  • Corollaries specialize the general bound to Bayesian transfer learning and to distribution shifts constrained within epsilon-neighborhoods, providing more targeted guidance for those scenarios.

Abstract

Uncertainty-aware machine learners, such as Bayesian neural networks, output a quantification of uncertainty instead of a point prediction. We provide uncertainty-aware learners with a principled framework to characterize, and identify ways to eliminate, errors that arise from reducible (epistemic) uncertainty. We introduce a principled definition of epistemic error, and provide a decompositional epistemic error bound which operates in the very general setting of imperfect multitask learning under distribution shift. In this setting, the training (source) data may arise from multiple tasks, the test (target) data may differ systematically from the source data tasks, and/or the learner may not arrive at an accurate characterization of the source data. Our bound separately attributes epistemic errors to each of multiple aspects of the learning procedure and environment. As corollaries of the general result, we provide epistemic error bounds specialized to the settings of Bayesian transfer learning and distribution shift within \epsilon-neighborhoods.