The Diminishing Returns of Early-Exit Decoding in Modern LLMs

arXiv cs.CL / 3/26/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper re-evaluates early-exit decoding in modern LLMs, arguing that newer training recipes and architectures may have less layer redundancy, reducing early-exit opportunities.
  • It introduces an “intrinsic suitability” metric and a benchmark to measure and compare early-exit benefits across models and workloads.
  • The authors find a diminishing trend in early-exit effectiveness across newer model generations, suggesting fewer gains from stopping early as models evolve.
  • The study reports that dense transformer models generally have more early-exit potential than Mixture-of-Experts and State Space Models.
  • It also finds that larger models (especially those above ~20B parameters) and base pretrained models without specialized tuning tend to show higher early-exit potential.

Abstract

In Large Language Model (LLM) inference, early-exit refers to stopping computation at an intermediate layer once the prediction is sufficiently confident, thereby reducing latency and cost. However, recent LLMs adopt improved pretraining recipes and architectures that reduce layer redundancy, potentially limiting early-exit opportunities. We re-evaluate layer-wise early-exit in modern LLMs and analyze how intermediate representations evolve during training. We introduce a metric to quantify a model's intrinsic suitability for early-exit and propose a benchmark for researchers to explore the potential early-exit benefits on different models and workloads. Our results show a diminishing trend in early-exit effectiveness across newer model generations. We further find that dense transformers generally offer greater early-exit potential than Mixture-of-Experts and State Space Models. In addition, larger models, particularly those with more than 20 billion parameters, and base pretrained models without specialized tuning tend to exhibit higher early-exit potential.