Timestep-Aware Block Masking for Efficient Diffusion Model Inference
arXiv cs.CV / 3/23/2026
📰 NewsDeveloper Stack & InfrastructureTools & Practical UsageModels & Research
Key Points
- Introduces timestep-specific masking to reduce inference latency in diffusion probabilistic models by deciding which blocks to execute or bypass at each denoising step.
- Learns masks per timestep to enable feature reuse and avoid full-chain backpropagation, yielding memory-efficient training.
- Adds a timestep-aware loss scaling and a knowledge-guided mask rectification strategy to preserve fidelity during sensitive denoising phases and prune redundant dependencies.
- Demonstrates architecture-agnostic applicability and reports efficiency gains across DDPM, LDM, DiT, and PixArt, with code to be released.