Timestep-Aware Block Masking for Efficient Diffusion Model Inference

arXiv cs.CV / 3/23/2026

📰 NewsDeveloper Stack & InfrastructureTools & Practical UsageModels & Research

Key Points

  • Introduces timestep-specific masking to reduce inference latency in diffusion probabilistic models by deciding which blocks to execute or bypass at each denoising step.
  • Learns masks per timestep to enable feature reuse and avoid full-chain backpropagation, yielding memory-efficient training.
  • Adds a timestep-aware loss scaling and a knowledge-guided mask rectification strategy to preserve fidelity during sensitive denoising phases and prune redundant dependencies.
  • Demonstrates architecture-agnostic applicability and reports efficiency gains across DDPM, LDM, DiT, and PixArt, with code to be released.

Abstract

Diffusion Probabilistic Models (DPMs) have achieved great success in image generation but suffer from high inference latency due to their iterative denoising nature. Motivated by the evolving feature dynamics across the denoising trajectory, we propose a novel framework to optimize the computational graph of pre-trained DPMs on a per-timestep basis. By learning timestep-specific masks, our method dynamically determines which blocks to execute or bypass through feature reuse at each inference stage. Unlike global optimization methods that incur prohibitive memory costs via full-chain backpropagation, our method optimizes masks for each timestep independently, ensuring a memory-efficient training process. To guide this process, we introduce a timestep-aware loss scaling mechanism that prioritizes feature fidelity during sensitive denoising phases, complemented by a knowledge-guided mask rectification strategy to prune redundant spatial-temporal dependencies. Our approach is architecture-agnostic and demonstrates significant efficiency gains across a broad spectrum of models, including DDPM, LDM, DiT, and PixArt. Experimental results show that by treating the denoising process as a sequence of optimized computational paths, our method achieves a superior balance between sampling speed and generative quality. Our code will be released.