FastDiSS: Few-step Match Many-step Diffusion Language Model on Sequence-to-Sequence Generation--Full Version

arXiv cs.CL / 4/8/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that self-conditioning in continuous diffusion language models breaks down under few-step sampling, where inaccurate prior estimates compound across denoising steps and significantly degrade output quality.
  • It proposes FastDiSS, a training framework that perturbs the self-conditioning signal to better match the noise characteristics seen during inference, reducing the approximation gap.
  • FastDiSS also adds a token-level noise-awareness mechanism to avoid training saturation and improve optimization stability.
  • Experiments on conditional generation benchmarks show that FastDiSS improves over standard continuous diffusion models and can deliver up to 400x faster inference while remaining competitive with one-step diffusion approaches.

Abstract

Self-conditioning has been central to the success of continuous diffusion language models, as it allows models to correct previous errors. Yet its ability degrades precisely in the regime where diffusion is most attractive for deployment: few-step sampling for fast inference. In this study, we show that when models only have a few denoising steps, inaccurate self-conditioning induces a substantial approximation gap; this mistake compounds across denoising steps and ultimately dominate the sample quality. To address this, we propose a novel training framework that handles these errors during learning by perturbing the self-conditioning signal to match inference noise, improving robustness to prior estimation errors. In addition, we introduce a token-level noise-awareness mechanism that prevents training from saturation, hence improving optimization. Extensive experiments across conditional generation benchmarks demonstrate that our framework surpasses standard continuous diffusion models while providing up to 400x faster inference speed, and remains competitive against other one-step diffusion frameworks.