InstaVSR: Taming Diffusion for Efficient and Temporally Consistent Video Super-Resolution

arXiv cs.CV / 3/30/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces InstaVSR, a diffusion-based framework aimed at efficient and temporally consistent video super-resolution from low-resolution inputs.
  • It tackles video diffusion’s two key issues—temporal instability from strong generative priors and high multi-frame diffusion compute costs—via a lightweight, pruned one-step diffusion backbone.
  • InstaVSR improves frame-to-frame consistency using recurrent training with flow-guided temporal regularization.
  • To maintain perceptual quality despite simplifying the backbone, it applies dual-space adversarial learning in both latent and pixel domains.
  • The authors report strong efficiency results: on an NVIDIA RTX 4090, it super-resolves a 30-frame, 2K×2K video in under one minute using about 7 GB of memory while producing smoother temporal transitions than prior diffusion VSR approaches.

Abstract

Video super-resolution (VSR) seeks to reconstruct high-resolution frames from low-resolution inputs. While diffusion-based methods have substantially improved perceptual quality, extending them to video remains challenging for two reasons: strong generative priors can introduce temporal instability, and multi-frame diffusion pipelines are often too expensive for practical deployment. To address both challenges simultaneously, we propose InstaVSR, a lightweight diffusion framework for efficient video super-resolution. InstaVSR combines three ingredients: (1) a pruned one-step diffusion backbone that removes several costly components from conventional diffusion-based VSR pipelines, (2) recurrent training with flow-guided temporal regularization to improve frame-to-frame stability, and (3) dual-space adversarial learning in latent and pixel spaces to preserve perceptual quality after backbone simplification. On an NVIDIA RTX 4090, InstaVSR processes a 30-frame video at 2K\times2K resolution in under one minute with only 7 GB of memory usage, substantially reducing the computational cost compared to existing diffusion-based methods while maintaining favorable perceptual quality with significantly smoother temporal transitions.