Calibrated Speculative Decoding: Frequency-Guided Candidate Selection for Efficient Inference

arXiv cs.CL / 4/16/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes Calibrated Speculative Decoding (CSD) to reduce speculative decoding’s false rejections caused by lexically divergent but semantically correct draft tokens.
  • CSD is a training-free approach that uses Frequency-Guided Candidate Selection and Probability-Guarded Acceptance, with two lightweight modules: Online Correction Memory for recurring divergence rescue candidates and Semantic Consistency Gating based on probability ratios.
  • Experiments across multiple large language models show CSD improves inference throughput, with a reported peak speedup of 2.33x.
  • The method maintains accuracy across tasks while providing additional performance gains on complex reasoning datasets, positioning it as a practical, lightweight upgrade for LLM deployments.

Abstract

Speculative decoding accelerates autoregressive generation by letting draft tokens bypass full verification, but conventional frameworks suffer from frequent false rejections, particularly when draft models produce semantically correct but lexically divergent outputs. In this paper, we present Calibrated Speculative Decoding (CSD), a training-free framework that recovers valid tokens discarded by standard verification. Guided by the principle of "Frequency-Guided Candidate Selection and Probability-Guarded Acceptance," CSD incorporates two lightweight modules: Online Correction Memory, which aggregates historical rejections to propose recurring divergence patterns as rescue candidates, and Semantic Consistency Gating, which verifies candidate admissibility using probability ratios instead of exact token matching. Our evaluation across diverse large language models demonstrates that CSD outperforms existing methods, achieving a peak throughput speedup of 2.33x. CSD preserves model accuracy across all tasks while further boosting performance on complex reasoning datasets. These results establish CSD as a highly effective, lightweight solution for practical LLM deployments.