ReDiffuse: Rotation Equivariant Diffusion Model for Multi-focus Image Fusion

arXiv cs.CV / 3/24/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • この論文は、マルチフォーカス画像融合(MFIF)において、ディフュージョンモデル適用時に生じる“ワープ/変形”由来のアーティファクト問題に対し、回転同変性を埋め込むことが重要だと述べています。
  • 提案手法ReDiffuseは、拡散モデルの基本アーキテクチャを設計し、エンドツーエンドで回転同変性を達成できるように構築しています。
  • 回転同変性構造の妥当性を支えるため、内在する“同変性誤差”を評価する理論解析も行っています。
  • Lytro, MFFW, MFI-WHU, Road-MFの4データセットで複数手法と比較し、6つの評価指標で0.28〜6.64%の改善を示しており、競争力のある性能を報告しています。
  • 実装コードはGitHubで公開されており、再現・追試や今後の拡張が可能です。

Abstract

Diffusion models have achieved impressive performance on multi-focus image fusion (MFIF). However, a key challenge in applying diffusion models to the ill-posed MFIF problem is that defocus blur can make common symmetric geometric structures (e.g., textures and edges) appear warped and deformed, often leading to unexpected artifacts in the fused images. Therefore, embedding rotation equivariance into diffusion networks is essential, as it enables the fusion results to faithfully preserve the original orientation and structural consistency of geometric patterns underlying the input images. Motivated by this, we propose ReDiffuse, a rotation-equivariant diffusion model for MFIF. Specifically, we carefully construct the basic diffusion architectures to achieve end-to-end rotation equivariance. We also provide a rigorous theoretical analysis to evaluate its intrinsic equivariance error, demonstrating the validity of embedding equivariance structures. ReDiffuse is comprehensively evaluated against various MFIF methods across four datasets (Lytro, MFFW, MFI-WHU, and Road-MF). Results demonstrate that ReDiffuse achieves competitive performance, with improvements of 0.28-6.64\% across six evaluation metrics. The code is available at https://github.com/MorvanLi/ReDiffuse.

ReDiffuse: Rotation Equivariant Diffusion Model for Multi-focus Image Fusion | AI Navigate