Uncertainty Quantification in CNN Through the Bootstrap of Convex Neural Networks

arXiv cs.LG / 4/15/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that uncertainty quantification (UQ) for convolutional neural networks (CNNs) has been underdeveloped, limiting high-stakes domains like medicine where prediction uncertainty matters.
  • It proposes a bootstrap-based framework that uses “convexified neural networks” to provide theoretical consistency guarantees for the quality of estimated uncertainty.
  • The method is designed to be computationally efficient by using warm-starts in each bootstrap iteration rather than refitting the model from scratch.
  • The authors introduce a transfer learning approach to extend the framework beyond convexified networks so it can work with arbitrary neural network architectures.
  • Experiments on multiple image datasets show improved performance versus baseline CNN approaches and several state-of-the-art UQ methods.

Abstract

Despite the popularity of Convolutional Neural Networks (CNN), the problem of uncertainty quantification (UQ) of CNN has been largely overlooked. Lack of efficient UQ tools severely limits the application of CNN in certain areas, such as medicine, where prediction uncertainty is critically important. Among the few existing UQ approaches that have been proposed for deep learning, none of them has theoretical consistency that can guarantee the uncertainty quality. To address this issue, we propose a novel bootstrap based framework for the estimation of prediction uncertainty. The inference procedure we use relies on convexified neural networks to establish the theoretical consistency of bootstrap. Our approach has a significantly less computational load than its competitors, as it relies on warm-starts at each bootstrap that avoids refitting the model from scratch. We further explore a novel transfer learning method so our framework can work on arbitrary neural networks. We experimentally demonstrate our approach has a much better performance compared to other baseline CNNs and state-of-the-art methods on various image datasets.