AgentFoX: LLM Agent-Guided Fusion with eXplainability for AI-Generated Image Detection

arXiv cs.CV / 3/25/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • AgentFoXは、AI生成画像検出を単一モデルの判定ではなく、LLMエージェントが複数フェーズで分析する動的プロセスとして再定義するフレームワークです。
  • 事前に用意した知識ベース(calibrated Expert Profiles と contextual Clustering Profiles)に基づき、意味的評価から信号レベルの専門家エビデンス統合へ段階的に進み、矛盾は構造化推論で解消します。
  • 既存手法が周波数領域パターンや意味的不整合など特定の人工物に最適化されがちな点を、検出根拠を統合して改善しようとしています。
  • 検出結果を単なる2値(合成/本物)ではなく、人が読める詳細なフォレンジックレポートとして提示し、解釈可能性と信頼性の向上を狙います。
  • 将来の新しいフォレンジック手法を追加しながら拡張できるスケーラブルなエージェント的統合パラダイムも提案しています。

Abstract

The increasing realism of AI-Generated Images (AIGI) has created an urgent need for forensic tools capable of reliably distinguishing synthetic content from authentic imagery. Existing detectors are typically tailored to specific forgery artifacts--such as frequency-domain patterns or semantic inconsistencies--leading to specialized performance and, at times, conflicting judgments. To address these limitations, we present \textbf{AgentFoX}, a Large Language Model-driven framework that redefines AIGI detection as a dynamic, multi-phase analytical process. Our approach employs a quick-integration fusion mechanism guided by a curated knowledge base comprising calibrated Expert Profiles and contextual Clustering Profiles. During inference, the agent begins with high-level semantic assessment, then transitions to fine-grained, context-aware synthesis of signal-level expert evidence, resolving contradictions through structured reasoning. Instead of returning a coarse binary output, AgentFoX produces a detailed, human-readable forensic report that substantiates its verdict, enhancing interpretability and trustworthiness for real-world deployment. Beyond providing a novel detection solution, this work introduces a scalable agentic paradigm that facilitates intelligent integration of future and evolving forensic tools.