TAB-AUDIT: Detecting AI-Fabricated Scientific Tables via Multi-View Likelihood Mismatch

arXiv cs.CL / 3/23/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper TAB-AUDIT investigates detection of AI-generated fabricated scientific tables in empirical NLP papers and introduces the FabTab benchmark with 1,173 AI-generated and 1,215 human-authored papers.
  • It identifies discriminative features, notably within-table mismatch, which captures the perplexity gap between a table's skeleton and its numerical content, to distinguish fabricated tables.
  • A RandomForest model using these features significantly outperforms prior methods, achieving 0.987 AUROC in-domain and 0.883 AUROC out-of-domain.
  • The findings position experimental tables as a critical forensic signal for detecting AI-generated scientific fraud and establish a new benchmark for future research.

Abstract

AI-generated fabricated scientific manuscripts raise growing concerns with large-scale breaches of academic integrity. In this work, we present the first systematic study on detecting AI-generated fabricated scientific tables in empirical NLP papers, as information in tables serve as critical evidence for claims. We construct FabTab, the first benchmark dataset of fabricated manuscripts with tables, comprising 1,173 AI-generated papers and 1,215 human-authored ones in empirical NLP. Through a comprehensive analysis, we identify systematic differences between fabricated and real tables and operationalize them into a set of discriminative features within the TAB-AUDIT framework. The key feature, within-table mismatch, captures the perplexity gap between a table's skeleton and its numerical content. Experimental results show that RandomForest built on these features significantly outperform prior state-of-the-art methods, achieving 0.987 AUROC in-domain and 0.883 AUROC out-of-domain. Our findings highlight experimental tables as a critical forensic signal for detecting AI-generated scientific fraud and provide a new benchmark for future research.