EvaNet: Towards More Efficient and Consistent Infrared and Visible Image Fusion Assessment

arXiv cs.CV / 4/6/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that common image-fusion evaluation metrics are often borrowed from other vision tasks, leading to poor quality measurement and heavy computation costs.
  • It introduces EvaNet, a unified, lightweight learning-based evaluation framework that first decomposes a fused image into infrared and visible components and then evaluates information preservation for each.
  • Training uses contrastive learning and incorporates perceptual scene assessment guidance from a large language model to better align the evaluation model with human-like perception.
  • The work also proposes a consistency evaluation approach that measures agreement between fusion metrics and human visual perception via no-reference scores and downstream task performance.
  • Experiments report substantially improved efficiency (up to 1,000× faster) and higher consistency across standard image-fusion benchmarks, with code planned for public release.

Abstract

Evaluation is essential in image fusion research, yet most existing metrics are directly borrowed from other vision tasks without proper adaptation. These traditional metrics, often based on complex image transformations, not only fail to capture the true quality of the fusion results but also are computationally demanding. To address these issues, we propose a unified evaluation framework specifically tailored for image fusion. At its core is a lightweight network designed efficiently to approximate widely used metrics, following a divide-and-conquer strategy. Unlike conventional approaches that directly assess similarity between fused and source images, we first decompose the fusion result into infrared and visible components. The evaluation model is then used to measure the degree of information preservation in these separated components, effectively disentangling the fusion evaluation process. During training, we incorporate a contrastive learning strategy and inform our evaluation model by perceptual scene assessment provided by a large language model. Last, we propose the first consistency evaluation framework, which measures the alignment between image fusion metrics and human visual perception, using both independent no-reference scores and downstream tasks performance as objective references. Extensive experiments show that our learning-based evaluation paradigm delivers both superior efficiency (up to 1,000 times faster) and greater consistency across a range of standard image fusion benchmarks. Our code will be publicly available at https://github.com/AWCXV/EvaNet.