AI Navigate

UCAN: Unified Convolutional Attention Network for Expansive Receptive Fields in Lightweight Super-Resolution

arXiv cs.CV / 3/13/2026

📰 NewsModels & Research

Key Points

  • UCAN is a lightweight network that unifies convolution and attention to expand the effective receptive field efficiently, enabling high-quality image super-resolution on resource-constrained devices.
  • It combines window-based spatial attention with a Hedgehog Attention mechanism to model both local texture and long-range dependencies.
  • A distillation-based large-kernel module preserves high-frequency structure without heavy computation, and cross-layer parameter sharing further reduces model complexity.
  • Empirical results show UCAN-L achieving 31.63 dB PSNR on Manga109 (4x) with 48.4G MACs and 27.79 dB on BSDS100, outperforming recent lightweight models and highlighting a favorable accuracy-efficiency trade-off.

Abstract

Hybrid CNN-Transformer architectures achieve strong results in image super-resolution, but scaling attention windows or convolution kernels significantly increases computational cost, limiting deployment on resource-constrained devices. We present UCAN, a lightweight network that unifies convolution and attention to expand the effective receptive field efficiently. UCAN combines window-based spatial attention with a Hedgehog Attention mechanism to model both local texture and long-range dependencies, and introduces a distillation-based large-kernel module to preserve high-frequency structure without heavy computation. In addition, we employ cross-layer parameter sharing to further reduce complexity. On Manga109 (4\times), UCAN-L achieves 31.63 dB PSNR with only 48.4G MACs, surpassing recent lightweight models. On BSDS100, UCAN attains 27.79 dB, outperforming methods with significantly larger models. Extensive experiments show that UCAN achieves a superior trade-off between accuracy, efficiency, and scalability, making it well-suited for practical high-resolution image restoration.