Multimodal Claim Extraction for Fact-Checking

arXiv cs.CL / 4/21/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that automated fact-checking needs claim extraction methods that account for multimodal misinformation, since real-world posts often pair informal text with images like memes and screenshots.
  • It introduces what it claims is the first benchmark for multimodal claim extraction from social media, with posts (text plus one or more images) annotated using gold-standard claims from professional fact-checkers.
  • The authors evaluate current state-of-the-art multimodal LLMs using a three-part framework covering semantic alignment, faithfulness, and decontextualization, finding that baseline models struggle with rhetorical intent and contextual cues.
  • To improve performance, the paper proposes MICE, an intent-aware framework designed to better handle intent-critical cases and deliver measurable gains.
  • Overall, the work combines a new dataset/benchmark, an evaluation methodology, and a targeted model framework aimed at making multimodal fact-checking more reliable.

Abstract

Automated Fact-Checking (AFC) relies on claim extraction as a first step, yet existing methods largely overlook the multimodal nature of today's misinformation. Social media posts often combine short, informal text with images such as memes, screenshots, and photos, creating challenges that differ from both text-only claim extraction and well-studied multimodal tasks like image captioning or visual question answering. In this work, we present the first benchmark for multimodal claim extraction from social media, consisting of posts containing text and one or more images, annotated with gold-standard claims derived from real-world fact-checkers. We evaluate state-of-the-art multimodal LLMs (MLLMs) under a three-part evaluation framework (semantic alignment, faithfulness, and decontextualization) and find that baseline MLLMs struggle to model rhetorical intent and contextual cues. To address this, we introduce MICE, an intent-aware framework which shows improvements in intent-critical cases.

Multimodal Claim Extraction for Fact-Checking | AI Navigate