VARestorer: One-Step VAR Distillation for Real-World Image Super-Resolution

arXiv cs.CV / 4/24/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces VARestorer, a distillation method that converts a pre-trained text-to-image visual autoregressive (VAR) model into a one-step model for real-world image super-resolution (Real-ISR).
  • It addresses ISR-specific problems where causal, next-scale prediction and iterative autoregressive refinement lead to blurry outputs and coherence degradation due to global context underuse and error accumulation.
  • VARestorer avoids iterative refinement by using distribution matching, which reduces error propagation and substantially lowers inference time.
  • The approach adds pyramid image conditioning with cross-scale attention to enable bidirectional interactions across scales, ensuring later low-quality (LQ) tokens are not neglected by the transformer.
  • Experiments report state-of-the-art results on DIV2K (72.32 MUSIQ and 0.7669 CLIPIQA) and 10× faster inference versus conventional VAR-based inference, while fine-tuning only 1.2% of parameters via parameter-efficient adapters.

Abstract

Recent advancements in visual autoregressive models (VAR) have demonstrated their effectiveness in image generation, highlighting their potential for real-world image super-resolution (Real-ISR). However, adapting VAR for ISR presents critical challenges. The next-scale prediction mechanism, constrained by causal attention, fails to fully exploit global low-quality (LQ) context, resulting in blurry and inconsistent high-quality (HQ) outputs. Additionally, error accumulation in the iterative prediction severely degrades coherence in ISR task. To address these issues, we propose VARestorer, a simple yet effective distillation framework that transforms a pre-trained text-to-image VAR model into a one-step ISR model. By leveraging distribution matching, our method eliminates the need for iterative refinement, significantly reducing error propagation and inference time. Furthermore, we introduce pyramid image conditioning with cross-scale attention, which enables bidirectional scale-wise interactions and fully utilizes the input image information while adapting to the autoregressive mechanism. This prevents later LQ tokens from being overlooked in the transformer. By fine-tuning only 1.2\% of the model parameters through parameter-efficient adapters, our method maintains the expressive power of the original VAR model while significantly enhancing efficiency. Extensive experiments show that VARestorer achieves state-of-the-art performance with 72.32 MUSIQ and 0.7669 CLIPIQA on DIV2K dataset, while accelerating inference by 10 times compared to conventional VAR inference.