arXiv implements 1-year ban for papers containing incontrovertible evidence of unchecked LLM-generated errors, such as hallucinated references or results. [N]

Reddit r/MachineLearning / 5/15/2026

📰 NewsSignals & Early TrendsIndustry & Market Moves

Key Points

  • arXivのモデレーターは、LLM生成の結果が十分に検証されていないことを示す「疑いを覆しえない証拠」がある投稿に対して、罰則を明確化しました。
  • 罰則は、該当する論文の投稿者に対するarXivでの1年間の出禁で、その後のarXiv投稿には信頼できる査読付き学術誌での事前受理が必要になります。
  • 「疑いを覆しえない証拠」の例として、存在しない参考文献(ハルシネーション)、LLMによるメタコメント(要約作成の案内や変更提案)、表の数値などを実験結果ではなく例示のままにしているような指示が挙げられています。
  • arXivは、生成AIツールが不適切な表現、剽窃、偏り、誤り、誤った参照、誤解を招く内容を含む出力を作った場合でも、それを科学的成果物に含めた著者側に責任があるという立場を強調しています。

From Thomas G. Dietterich (arXiv moderator for cs.LG) on 𝕏 (thread):
https://x.com/tdietterich/status/2055000956144935055
https://xcancel.com/tdietterich/status/2055000956144935055

"Attention arXiv authors: Our Code of Conduct states that by signing your name as an author of a paper, each author takes full responsibility for all its contents, irrespective of how the contents were generated.

If generative AI tools generate inappropriate language, plagiarized content, biased content, errors, mistakes, incorrect references, or misleading content, and that output is included in scientific works, it is the responsibility of the author(s).

We have recently clarified our penalties for this. If a submission contains incontrovertible evidence that the authors did not check the results of LLM generation, this means we can't trust anything in the paper.

The penalty is a 1-year ban from arXiv followed by the requirement that subsequent arXiv submissions must first be accepted at a reputable peer-reviewed venue.

Examples of incontrovertible evidence: hallucinated references, meta-comments from the LLM ("here is a 200 word summary; would you like me to make any changes?"; "the data in this table is illustrative, fill it in with the real numbers from your experiments")."

submitted by /u/Nunki08
[link] [comments]