Multi-Feature Fusion Approach for Generative AI Images Detection

arXiv cs.CV / 4/1/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses the growing challenge of detecting GenAI-generated images as generative models produce increasingly realistic synthetic photos.
  • It proposes a multi-feature fusion detector that combines complementary cues from MSCN (low-level statistical deviations), CLIP embeddings (high-level semantic coherence), and MLBP (mid-level texture anomalies).
  • Experiments across four benchmark datasets show that relying on any single feature space leads to unstable performance across different generative models.
  • The combined fusion of all three representations delivers more consistent, higher detection accuracy, especially in difficult mixed-model evaluation scenarios.
  • Compared with existing state-of-the-art approaches, the proposed hybrid framework improves performance across all evaluated datasets while offering a generalizable method for integrating visual cues.

Abstract

The rapid evolution of Generative AI (GenAI) models has led to synthetic images of unprecedented realism, challenging traditional methods for distinguishing them from natural photographs. While existing detectors often rely on single-feature spaces, such as statistical regularities, semantic embeddings, or texture patterns, these approaches tend to lack robustness when confronted with diverse and evolving generative models. In this work, we investigate and systematically evaluate a multi-feature fusion framework that combines complementary cues from three distinct spaces: (1) Mean Subtracted Contrast Normalized (MSCN) features capturing low-level statistical deviations; (2) CLIP embeddings encoding high-level semantic coherence; and (3) Multi-scale Local Binary Patterns (MLBP) characterizing mid-level texture anomalies. Through extensive experiments on four benchmark datasets covering a wide range of generative models, we show that individual feature spaces exhibit significant performance variability across different generators. Crucially, the fusion of all three representations yields superior and more consistent performance, particularly in a challenging mixed-model scenario. Compared to state-of-the-art methods, the proposed framework yields consistently improved performance across all evaluated datasets. Overall, this work highlights the importance of hybrid representations for robust GenAI image detection and provides a principled framework for integrating complementary visual cues.

Multi-Feature Fusion Approach for Generative AI Images Detection | AI Navigate