ZEUS: Accelerating Diffusion Models with Only Second-Order Predictor

arXiv cs.LG / 4/3/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • Diffusion-based denoising generators can be slow because sampling requires many iterative denoiser calls, motivating training-free methods to accelerate inference by step skipping or sparsification.
  • The paper argues that existing aggressive training-free accelerators often rely on higher-order predictors that amplify error, and that architectural changes can complicate deployment.
  • It introduces ZEUS, which uses a second-order predictor plus an interleaved scheme to stabilize consecutive skipping and avoid back-to-back extrapolation failures.
  • ZEUS is designed to add essentially zero overhead, without feature caches or architectural modifications, and it remains compatible across different model backbones, objectives, and solvers.
  • Experiments on image and video generation show up to 3.2× end-to-end speedup while preserving perceptual quality, improving over recent training-free baselines, with code released on GitHub.

Abstract

Denoising generative models deliver high-fidelity generation but remain bottlenecked by inference latency due to the many iterative denoiser calls required during sampling. Training-free acceleration methods reduce latency by either sparsifying the model architecture or shortening the sampling trajectory. Current training-free acceleration methods are more complex than necessary: higher-order predictors amplify error under aggressive speedups, and architectural modifications hinder deployment. Beyond 2x acceleration, step skipping creates structural scarcity -- at most one fresh evaluation per local window -- leaving the computed output and its backward difference as the only causally grounded information. Based on this, we propose ZEUS, an acceleration method that predicts reduced denoiser evaluations using a second-order predictor, and stabilizes aggressive consecutive skipping with an interleaved scheme that avoids back-to-back extrapolations. ZEUS adds essentially zero overhead, no feature caches, and no architectural modifications, and it is compatible with different backbones, prediction objectives, and solver choices. Across image and video generation, ZEUS consistently improves the speed-fidelity performance over recent training-free baselines, achieving up to 3.2x end-to-end speedup while maintaining perceptual quality. Our code is available at: https://github.com/Ting-Justin-Jiang/ZEUS.