Breaking Block Boundaries: Anchor-based History-stable Decoding for Diffusion Large Language Models
arXiv cs.CL / 4/13/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that semi-autoregressive decoding in diffusion LLMs has inherent block constraints that delay decoding of many cross-block stable tokens.
- It presents three observations about stable token identification: naive lookahead is unreliable, token stability correlates with convergence trends, and historical information becomes isolated across blocks.
- To address this, the authors propose Anchor-based History-stable Decoding (AHD), a training-free, plug-and-play dynamic decoding method that tracks token stability via dynamic anchors in real time.
- Experiments across language, vision-language, and audio-language tasks show AHD improves both performance and inference efficiency and can reverse the typical degradation seen in other advanced acceleration strategies.
- On the BBH benchmark specifically, AHD reduces decoding steps by 80% while improving performance by 3.67%.
Related Articles

Black Hat Asia
AI Business

Apple is building smart glasses without a display to serve as an AI wearable
THE DECODER

Why Fashion Trend Prediction Isn’t Enough Without Generative AI
Dev.to
Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to
Chatbot vs Voicebot: The Real Business Decision Nobody Talks About
Dev.to