A Comparative Analysis on the Performance of Upper Confidence Bound Algorithms in Adaptive Deep Neural Networks

arXiv cs.LG / 4/29/2026

📰 NewsModels & Research

Key Points

  • The paper targets edge computing, where strict latency and energy limits require adaptive neural-network inference that balances cost against predictive accuracy.
  • Building on ADNNs that use a Multi-Armed Bandit approach, it expands beyond the commonly used UCB1 strategy by adding four UCB variants: UCB-V, UCB-Tuned, UCB-Bayes, and UCB-BwK.
  • The authors run a first comparative study of these UCB strategies, evaluating accuracy/latency/energy trade-offs using ResNet and MobileViT on CIFAR-10, CIFAR-10.1, and CIFAR-100.
  • All tested strategies show sub-linear cumulative regret, with UCB-Bayes converging fastest, and UCB-V and UCB-Tuned producing the best Pareto-optimal accuracy–latency and accuracy–energy results.

Abstract

Edge computing environments impose strict constraints on energy consumption and latency, making the deployment of deep neural networks a significant challenge. Therefore, smart and adaptive inference strategies that dynamically balance computational cost or latency with predictive accuracy are critical in edge computing scenarios. In this work, we build on Adaptive Deep Neural Networks (ADNNs) that employ the Multi-Armed Bandit (MAB) framework. Current literature leverages the first version of the Upper Confidence Bound (UCB1) strategy to dynamically select the optimal confidence threshold, enabling efficient early exits without sacrificing accuracy. However, we introduce four additional Upper Confidence Bound strategies in ADNNs, namely UCB-V, UCB-Tuned, UCB-Bayes, and UCB-BwK, and perform, for the first time, a comparative study of these strategies with respect to trade-offs between accuracy, energy consumption, and latency. The proposed UCB strategies are employed on the ResNet and MobileViT neural networks, and are evaluated on the benchmark datasets of CIFAR-10, CIFAR-10.1, and CIFAR-100. Experimental results demonstrate that all strategies achieve sub-linear cumulative regret, with UCB-Bayes converging the fastest, followed by UCB-Tuned and UCB-V. Finally, UCB-V and UCB-Tuned dominate the Pareto Frontiers of accuracy-latency and accuracy-energy trade-offs.