Is Human-Like Text Liked by Humans? Multilingual Human Detection and Preference Against AI

arXiv cs.CL / 4/30/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The study investigates how accurately humans can detect whether text is AI-generated or human-written across multiple languages and domains.
  • Using 16 datasets in 9 languages and 9 domains with 19 annotators, the researchers find an average detection accuracy of 87.6%, which is higher than earlier reports and suggests different upper bounds.
  • The main discrepancies between human and machine text are attributed to factors such as concreteness, cultural nuances, and diversity.
  • Providing prompts that explicitly explain the differences can improve detection in over half of the cases, partially narrowing the human–AI gap.
  • The researchers also observe that people do not always prefer human-written text, especially when the origin of the text is unclear, and they release the dataset and labels on GitHub.

Abstract

Prior studies have shown that distinguishing text generated by Large Language Models (LLMs) from human-written one is highly challenging for humans, and often no better than random guessing. To verify the generalizability of this finding across languages and domains, we perform an extensive case study to identify the upper bound of human detection accuracy. Across 16 datasets covering 9 languages and 9 domains, 19 annotators achieved an average detection accuracy of 87.6%, thus challenging previous conclusions. We find that major gaps between human and machine text lie in concreteness, cultural nuances, and diversity. Prompting by explicitly explaining the distinctions in the prompts can partially bridge the gaps in over 50% of the cases. However, we also find that humans do not always prefer human-written text, particularly when they cannot clearly identify its source. We release our dataset, the human labels, and the annotator metadata at https://github.com/xnlp-lab/HumanEval-MGT.