Faster LLM Inference via Sequential Monte Carlo

arXiv cs.LG / 4/20/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Sequential Monte Carlo Speculative Decoding (SMC-SD) to speed up LLM inference by improving on speculative decoding, which normally suffers throughput loss when draft and target models diverge.
  • Instead of truncating or rejecting draft token blocks at the first mismatch, SMC-SD reweights and importance-resamples a population of draft particles, turning rejection into an approximate inference mechanism.
  • The method is designed to trade some exactness for speed while retaining theoretical bounds on the per-step approximation error.
  • Because LLM inference is often memory-bandwidth bound, SMC-SD leverages idle compute to parallelize verification as a vectorized, fixed-size operation without rollback.
  • Experiments report 2.36× speed-up over speculative decoding and 5.2× over autoregressive decoding, while staying within 3% of the target model’s accuracy across reasoning, instruction-following, and coding benchmarks.

Abstract

Speculative decoding (SD) accelerates language model inference by drafting tokens from a cheap proposal model and verifying them against an expensive target model via rejection sampling. Because rejection truncates the draft block at the first error, throughput degrades when draft and target diverge. Rather than rejecting draft tokens outright, we propose to reweight them. To this end, we introduce sequential Monte Carlo speculative decoding (SMC-SD), which replaces token-level rejection with importance-weighted resampling over a population of draft particles. SMC-SD is a principled approximate inference scheme that trades exactness for additional speed, while preserving theoretical bounds on its per-step approximation error. Because LLM inference is memory bandwidth-bound, the arithmetic needed to draft particles and to score them in parallel comes nearly for free -- SMC-SD uses idle compute to turn verification into a vectorized, fixed-size operation with no rollback. Empirically, SMC-SD achieves 2.36x speed-up over speculative decoding and a 5.2x speed-up over autoregressive decoding, while remaining within 3% of the target model's accuracy on reasoning, instruction-following, and coding benchmarks.