AI Navigate

Efficient Video Diffusion with Sparse Information Transmission for Video Compression

arXiv cs.CV / 3/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Diff-SIT is a diffusion-based video compression framework designed to maximize perceptual quality and temporal consistency at ultra-low bitrates.
  • It introduces the Sparse Temporal Encoding Module (STEM) to sparsely encode the original frame sequence into an information-rich intermediate sequence, enabling significant bitrate savings.
  • It also introduces the One-Step Video Diffusion with Frame Type Embedder (ODFTE), which processes the intermediate sequence with Frame Type Embedder guidance to adapt reconstruction for different frame types.
  • The approach achieves state-of-the-art perceptual quality and temporal consistency on several datasets, and the authors have released code at GitHub.

Abstract

Video compression aims to maximize reconstruction quality with minimal bitrates. Beyond standard distortion metrics, perceptual quality and temporal consistency are also critical. However, at ultra-low bitrates, traditional end-to-end compression models tend to produce blurry images of poor perceptual quality. Besides, existing generative compression methods often treat video frames independently and show limitations in time coherence and efficiency. To address these challenges, we propose the Efficient Video Diffusion with Sparse Information Transmission (Diff-SIT), which comprises the Sparse Temporal Encoding Module (STEM) and the One-Step Video Diffusion with Frame Type Embedder (ODFTE). The STEM sparsely encodes the original frame sequence into an information-rich intermediate sequence, achieving significant bitrate savings. Subsequently, the ODFTE processes this intermediate sequence as a whole, which exploits the temporal correlation. During this process, our proposed Frame Type Embedder (FTE) guides the diffusion model to perform adaptive reconstruction according to different frame types to optimize the overall quality. Extensive experiments on multiple datasets demonstrate that Diff-SIT establishes a new state-of-the-art in perceptual quality and temporal consistency, particularly in the challenging ultra-low-bitrate regime. Code is released at https://github.com/MingdeZhou/Diff-SIT.