Sparse VideoGen2: Accelerate Video Generation with Sparse Attention via Semantic-Aware Permutation

arXiv cs.CV / 5/4/2026

📰 NewsDeveloper Stack & InfrastructureSignals & Early TrendsModels & Research

Key Points

  • Diffusion Transformers for video generation face high latency because attention has quadratic complexity, motivating the use of sparse attention to speed up generation.
  • The paper finds that prior sparse attention methods underperform under the same compute budget due to poor critical-token selection (position-based rather than semantic) and inefficient GPU computation (critical tokens are scattered).
  • It introduces SVG2 (Sparse VideoGen2), a training-free framework that improves critical-token identification and reduces computation waste using a semantic-aware permutation based on k-means clustering and token reordering.
  • SVG2 further adds top-p dynamic budget control and customized kernel implementations, reporting up to 2.30× and 1.89× speedups while preserving video quality metrics (PSNR up to 30 on HunyuanVideo and 26 on Wan 2.1).
  • The authors open-source the codebase at the provided GitHub repository, enabling reproducibility and adoption by others working on efficient video generation.

Abstract

Diffusion Transformers (DiTs) are essential for video generation but suffer from significant latency due to the quadratic complexity of attention. By computing only critical tokens, sparse attention reduces computational costs and offers a promising acceleration approach. However, we identify that existing methods fail to approach optimal generation quality under the same computation budget for two reasons: (1) Inaccurate critical token identification: current methods cluster tokens based on position rather than semantics, leading to imprecise aggregated representations. (2) Excessive computation waste: critical tokens are scattered among non-critical ones, leading to wasted computation on GPUs, which are optimized for processing contiguous tokens. In this paper, we propose SVG2, a training-free framework that maximizes identification accuracy and minimizes computation waste, achieving a Pareto frontier trade-off between generation quality and efficiency. The core of SVG2 is semantic-aware permutation, which clusters and reorders tokens based on semantic similarity using k-means. This approach ensures both a precise cluster representation, improving identification accuracy, and a densified layout of critical tokens, enabling efficient computation without padding. Additionally, SVG2 integrates top-p dynamic budget control and customized kernel implementations, achieving up to 2.30x and 1.89x speedup while maintaining a PSNR of up to 30 and 26 on HunyuanVideo and Wan 2.1, respectively. Our code is open-sourced at \href{https://github.com/svg-project/Sparse-VideoGen}{https://github.com/svg-project/Sparse-VideoGen}.