Degradation-Robust Fusion: An Efficient Degradation-Aware Diffusion Framework for Multimodal Image Fusion in Arbitrary Degradation Scenarios

arXiv cs.CV / 4/13/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces “Degradation-Robust Fusion,” an efficient degradation-aware diffusion framework aimed at improving multimodal image fusion when inputs suffer from real-world degradations such as noise, blur, and low resolution.
  • It adapts diffusion approaches by using implicit denoising: instead of predicting diffusion noise explicitly, the model directly regresses the fused image to support flexible performance across varied degradation scenarios with limited sampling steps.
  • The method includes a joint observation-model correction mechanism that enforces both degradation consistency and fusion constraints during sampling to maintain high reconstruction accuracy.
  • Experiments across multiple fusion tasks and degradation configurations reportedly show the proposed approach outperforms existing methods, particularly under complex degradation conditions.

Abstract

Complex degradations like noise, blur, and low resolution are typical challenges in real world image fusion tasks, limiting the performance and practicality of existing methods. End to end neural network based approaches are generally simple to design and highly efficient in inference, but their black-box nature leads to limited interpretability. Diffusion based methods alleviate this to some extent by providing powerful generative priors and a more structured inference process. However, they are trained to learn a single domain target distribution, whereas fusion lacks natural fused data and relies on modeling complementary information from multiple sources, making diffusion hard to apply directly in practice. To address these challenges, this paper proposes an efficient degradation aware diffusion framework for image fusion under arbitrary degradation scenarios. Specifically, instead of explicitly predicting noise as in conventional diffusion models, our method performs implicit denoising by directly regressing the fused image, enabling flexible adaptation to diverse fusion tasks under complex degradations with limited steps. Moreover, we design a joint observation model correction mechanism that simultaneously imposes degradation and fusion constraints during sampling to ensure high reconstruction accuracy. Experiments on diverse fusion tasks and degradation configurations demonstrate the superiority of the proposed method under complex degradation scenarios.