Hero-Mamba: Mamba-based Dual Domain Learning for Underwater Image Enhancement

arXiv cs.CV / 4/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • Underwater images degrade due to light absorption and scattering, causing color distortion, low contrast, and blurred details that existing CNN and Transformer approaches struggle to handle efficiently.
  • Hero-Mamba introduces a Mamba-based dual-domain learning framework that processes both the spatial (RGB) domain and the spectral (FFT) domain in parallel to disentangle degradation factors.
  • The model’s Mamba-based SS2D blocks target long-range dependencies and global receptive fields with linear computational complexity, addressing Transformers’ quadratic cost for high-resolution inputs.
  • A ColorFusion block, guided by a background light prior, is designed to restore underwater color information with high fidelity.
  • Experiments on LSUI and UIEB show Hero-Mamba outperforming prior state-of-the-art methods, achieving PSNR 25.802 and SSIM 0.913 on LSUI.

Abstract

Underwater images often suffer from severe degradation, such as color distortion, low contrast, and blurred details, due to light absorption and scattering in water. While learning-based methods like CNNs and Transformers have shown promise, they face critical limitations: CNNs struggle to model the long-range dependencies needed for non-uniform degradation, and Transformers incur quadratic computational complexity, making them inefficient for high-resolution images. To address these challenges, we propose Hero-Mamba, a novel Mamba-based network that achieves efficient dual-domain learning for underwater image enhancement. Our approach uniquely processes information from both the spatial domain (RGB image) and the spectral domain (FFT components) in parallel. This dual-domain input allows the network to decouple degradation factors, separating color/brightness information from texture/noise. The core of our network utilizes Mamba-based SS2D blocks to capture global receptive fields and long-range dependencies with linear complexity, overcoming the limitations of both CNNs and Transformers. Furthermore, we introduce a ColorFusion block, guided by a background light prior, to restore color information with high fidelity. Extensive experiments on the LSUI and UIEB benchmark datasets demonstrate that Hero-Mamba outperforms state-of-the-art methods. Notably, our model achieves a PSNR of 25.802 and an SSIM of 0.913 on LSUI, validating its superior performance and generalization capabilities.