AI Navigate

MineDraft: A Framework for Batch Parallel Speculative Decoding

arXiv cs.AI / 3/20/2026

💬 OpinionDeveloper Stack & InfrastructureTools & Practical UsageModels & Research

Key Points

  • Speculative decoding accelerates LLM inference by using a smaller draft model to propose tokens that are later verified by a larger target model, but standard SD is limited by its strictly sequential drafting and verification stages.
  • MineDraft proposes a batch-parallel PSD framework that maintains two batches of requests and overlaps drafting for one batch with verification for the other to hide drafting latency.
  • Theoretical analysis shows PSD is substantially more efficient than standard SD.
  • Empirical results show significant improvements, with throughput up to 75% and end-to-end latency up to 39% faster, and MineDraft is implemented as a vLLM plugin to support production-ready inference systems.

Abstract

Speculative decoding (SD) accelerates large language model inference by using a smaller draft model to propose draft tokens that are subsequently verified by a larger target model. However, the performance of standard SD is often limited by the strictly sequential execution of these drafting and verification stages. To address this, this paper proposes MineDraft, a batch parallel speculative decoding (PSD) framework designed to effectively hide drafting latency by overlapping it with verification. Our theoretical analysis shows that PSD is substantially more efficient than standard SD. MineDraft realizes the PSD through a novel batch-parallel design that maintains two batches of requests, overlapping drafting for one batch with verification for the other. Our experimental results show significant improvements of MineDraft in both throughput (up to 75%) and end-to-end latency (up to 39%) over standard SD. Furthermore, we have implemented MineDraft as a plugin for vLLM, demonstrating its practicality for production-ready inference systems.