AgentPack: A Dataset of Code Changes, Co-Authored by Agents and Humans

arXiv cs.CL / 3/30/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • AgentPackは、Claude Code・OpenAI Codex・Cursor Agentが人間と共同で行った1.8M件のコード編集を集めたコーパスで、GitHub上の公開リポジトリ(2025年10月初旬まで)を対象にしています。
  • 従来の学習データはコミットメッセージの短さや複合的な変更、ボット由来のノイズなどで質に課題がありましたが、エージェント共同編集では意図や理由を自然言語でより明確に残しやすい点が利点として示されています。
  • 公開リポジトリでの取り込み段階ではメンテナが低品質な変更を除外することで、間接的な品質フィルタが働く可能性が論じられています。
  • 論文ではAgentPackの識別・キュレーション手順やエージェント採用トレンド、編集の構造的性質を分析し、AgentPackで微調整したモデルが従来の人間のみのコミット学習に比べて性能向上することを報告しています。

Abstract

Fine-tuning large language models for code editing has typically relied on mining commits and pull requests. The working hypothesis has been that commit messages describe human intent in natural language, and patches to code describe the changes that implement that intent. However, much of the previously collected data is noisy: commit messages are terse, human-written commits commingle several unrelated edits, and many commits come from simple, rule-based bots. The recent adoption of software engineering agents changes this landscape. Code changes \emph{co-authored} by humans and agents are often accompanied by substantially more explicit natural-language descriptions of intent and rationale. Moreover, when these changes land in public repositories, they are implicitly filtered by humans: maintainers discard low-quality commits to their projects. We present AgentPack, a corpus of 1.8M code edits co-authored by Claude Code, OpenAI Codex, and Cursor Agent across public GitHub projects up to early October 2025. We describe the identification and curation pipeline, quantify adoption trends of these agents, and analyze the structural properties of the edits. Finally, we show that models fine-tuned on AgentPack can outperform models trained on prior human-only commit corpora, highlighting the potential of using public data from software engineering agents to train future code-editing models.