AI Navigate

DART: Input-Difficulty-AwaRe Adaptive Threshold for Early-Exit DNNs

arXiv cs.AI / 3/16/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • DART introduces an input-difficulty-aware framework that quantifies input complexity with a lightweight estimator to guide early exits in DNNs, enabling adaptive inference on edge AI accelerators.
  • It features a joint exit policy optimization based on dynamic programming and an adaptive coefficient management system to balance accuracy and efficiency under varying workloads.
  • Experimental results on AlexNet, ResNet-18, and VGG-16 show up to 3.3x speedup, 5.1x energy reduction, and 42% lower average power with competitive accuracy; extending to Vision Transformers (LeViT) provides 5.0x power and 3.6x execution-time gains but may incur up to 17% accuracy loss.
  • The paper also introduces DAES (Difficulty-Aware Efficiency Score), a multi-objective metric that highlights improved accuracy, efficiency, and robustness versus baselines.

Abstract

Early-exit deep neural networks enable adaptive inference by terminating computation when sufficient confidence is achieved, reducing cost for edge AI accelerators in resource-constrained settings. Existing methods, however, rely on suboptimal exit policies, ignore input difficulty, and optimize thresholds independently. This paper introduces DART (Input-Difficulty-Aware Adaptive Threshold), a framework that overcomes these limitations. DART introduces three key innovations: (1) a lightweight difficulty estimation module that quantifies input complexity with minimal computational overhead, (2) a joint exit policy optimization algorithm based on dynamic programming, and (3) an adaptive coefficient management system. Experiments on diverse DNN benchmarks (AlexNet, ResNet-18, VGG-16) demonstrate that DART achieves up to \textbf{3.3\times} speedup, \textbf{5.1\times} lower energy, and up to \textbf{42\%} lower average power compared to static networks, while preserving competitive accuracy. Extending DART to Vision Transformers (LeViT) yields power (5.0\times) and execution-time (3.6\times) gains but also accuracy loss (up to 17 percent), underscoring the need for transformer-specific early-exit mechanisms. We further introduce the Difficulty-Aware Efficiency Score (DAES), a novel multi-objective metric, under which DART achieves up to a 14.8 improvement over baselines, highlighting superior accuracy, efficiency, and robustness trade-offs.