Focus on the Core: Empowering Diffusion Large Language Models by Self-Contrast

arXiv cs.CL / 5/5/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • Diffusion Large Language Models have strong global context modeling via iterative denoising, but existing decoding methods often act locally and ignore how information density varies across the context, hurting generation quality.
  • The paper identifies high-information-density (HD) tokens as a key factor: explicitly conditioning on HD tokens improves outputs, and HD tokens tend to be decoded earlier than neighboring tokens.
  • It proposes Focus on the Core (FoCore), a training-free decoding approach that remasks HD tokens as negative samples in a self-contrast scheme to better guide generation.
  • An accelerated variant, FoCore_A, detects when HD tokens converge and then runs parallel decoding on stable candidates within a local window to significantly reduce decoding time.
  • Experiments across math, code, and logical reasoning benchmarks show FoCore improves quality and FoCore_A improves efficiency for both LLaDA and Dream backbones; on HumanEval, pass@1 rises from 39.02 to 42.68, and latency drops from 20.76s to 8.64s (−58.4%).

Abstract

The iterative denoising paradigm of Diffusion Large Language Models (DLMs) endows them with a distinct advantage in global context modeling. However, current decoding strategies fail to leverage this capability, typically exhibiting a local preference that overlooks the heterogeneous information density within the context, ultimately degrading generation quality. To address this limitation, we systematically investigate high-information-density (HD) tokens and present two key findings: (1) explicitly conditioning on HD tokens substantially improves output quality; and (2) HD tokens exhibit an early-decoding tendency, converging earlier than surrounding tokens. Motivated by these findings, we propose Focus on the Core \textbf{(FoCore)}, a training-free decoding strategy that utilizes HD tokens in a self-contrast manner, wherein HD tokens are temporarily remasked as negative samples, to guide generation. We further introduce FoCore\_Accelerate \textbf{(FoCore\_A)}, an efficient variant that, upon detecting HD token convergence, performs parallel decoding over stable candidates within a local context window, substantially accelerating generation. Extensive experiments on math, code and logical reasoning benchmarks demonstrate that FoCore consistently improves generation quality and efficiency across both LLaDA and Dream backbones. For instance, on HumanEval, FoCore improves pass@1 from 39.02 to 42.68 over standard Classifier-Free Guidance, while FoCore-A reduces the number of decoding steps by 2.07x and per-sample latency from 20.76s to 8.64s (-58.4\%).