MGTEVAL: An Interactive Platform for Systemtic Evaluation of Machine-Generated Text Detectors

arXiv cs.CL / 4/29/2026

💬 OpinionDeveloper Stack & InfrastructureTools & Practical UsageModels & Research

Key Points

  • MGTEVAL is an extensible platform designed to enable systematic, reproducible evaluation of machine-generated text (MGT) detectors.
  • It addresses fragmentation in prior work by standardizing the evaluation workflow across dataset creation, text attacks, detector training, and performance measurement.
  • Users can build custom benchmarks by generating MGT with configurable LLMs, running 12 types of text attacks on test sets, and training detectors through a unified interface.
  • The platform reports multiple dimensions of results—effectiveness, robustness, and efficiency—and is accessible via both command-line and web interfaces.
  • By avoiding repeated code rewriting, MGTEVAL aims to make detector comparisons across datasets and settings more straightforward for researchers and practitioners.

Abstract

We present MGTEVAL, an extensible platform for systematic evaluation of Machine-Generated Text (MGT) detectors. Despite rapid progress in MGT detection, existing evaluations are often fragmented across datasets, preprocessing, attacks, and metrics, making results hard to compare and reproduce. MGTEVAL organizes the workflow into four components: Dataset Building, Dataset Attack, Detector Training, and Performance Evaluation. It supports constructing custom benchmarks by generating MGT with configurable LLMs, applying 12 text attacks to test sets, training detectors via a unified interface, and reporting effectiveness, robustness, and efficiency. The platform provides both command-line and Web-based interfaces for user-friendly experimentation without code rewriting.