Monte Carlo Stochastic Depth for Uncertainty Estimation in Deep Learning

arXiv cs.LG / 4/15/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses uncertainty quantification (UQ) for safety-critical deep neural network deployment, focusing on extending Monte Carlo–style Bayesian approximation beyond Monte Carlo Dropout to Stochastic Depth.
  • It provides theoretical connections between Monte Carlo Stochastic Depth (MCSD) and principled approximate variational Bayesian inference.
  • The authors run the first comprehensive benchmark of MCSD versus MCD and MC-DropBlock (MCDB) on state-of-the-art object detectors (YOLO, RT-DETR) using COCO and COCO-O.
  • Results show MCSD delivers highly competitive mAP while offering slight improvements in calibration (ECE) and uncertainty ranking (AUARC) compared with MCD, with strong computational efficiency.
  • Overall, the work positions MCSD as a theoretically grounded and empirically validated approach for efficient Bayesian approximation in modern architectures that rely on residual backbones.

Abstract

The deployment of deep neural networks in safety-critical systems necessitates reliable and efficient uncertainty quantification (UQ). A practical and widespread strategy for UQ is repurposing stochastic regularizers as scalable approximate Bayesian inference methods, such as Monte Carlo Dropout (MCD) and MC-DropBlock (MCDB). However, this paradigm remains under-explored for Stochastic Depth (SD), a regularizer integral to the residual-based backbones of most modern architectures. While prior work demonstrated its empirical promise for segmentation, a formal theoretical connection to Bayesian variational inference and a benchmark on complex, multi-task problems like object detection are missing. In this paper, we first provide theoretical insights connecting Monte Carlo Stochastic Depth (MCSD) to principled approximate variational inference. We then present the first comprehensive empirical benchmark of MCSD against MCD and MCDB on state-of-the-art detectors (YOLO, RT-DETR) using the COCO and COCO-O datasets. Our results position MCSD as a robust and computationally efficient method that achieves highly competitive predictive accuracy (mAP), notably yielding slight improvements in calibration (ECE) and uncertainty ranking (AUARC) compared to MCD. We thus establish MCSD as a theoretically-grounded and empirically-validated tool for efficient Bayesian approximation in modern deep learning.