FactAppeal: Identifying Epistemic Factual Appeals in News Media

arXiv cs.CL / 3/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Epistemic Appeal Identification, a task focused on determining whether and how news statements are made credible through external sources or evidence rather than just verifying claims.
  • It presents FactAppeal, a manually annotated dataset of 3,226 English news sentences with span-level labels for factual statements and the evidentiary sources they rely on.
  • The dataset includes fine-grained epistemic features such as source type (e.g., expert, witness, direct evidence), whether sources are named, and how attribution is expressed via direct/indirect quotation.
  • The authors evaluate multiple encoder and generative decoder models (2B–9B parameters) and report best performance using Gemma 2 9B with a macro-F1 score of 0.73.
  • The work reframes claim credibility as an interpretable structure of epistemic anchoring, enabling more nuanced NLP research on evidence-aware news understanding.

Abstract

How is a factual claim made credible? We propose the novel task of Epistemic Appeal Identification, which identifies whether and how factual statements have been anchored by external sources or evidence. To advance research on this task, we present FactAppeal, a manually annotated dataset of 3,226 English-language news sentences. Unlike prior resources that focus solely on claim detection and verification, FactAppeal identifies the nuanced epistemic structures and evidentiary basis underlying these claims and used to support them. FactAppeal contains span-level annotations which identify factual statements and mentions of sources on which they rely. Moreover, the annotations include fine-grained characteristics of factual appeals such as the type of source (e.g. Active Participant, Witness, Expert, Direct Evidence), whether it is mentioned by name, mentions of the source's role and epistemic credentials, attribution to the source via direct or indirect quotation, and other features. We model the task with a range of encoder models and generative decoder models in the 2B-9B parameter range. Our best performing model, based on Gemma 2 9B, achieves a macro-F1 score of 0.73.