SWE-chat: Coding Agent Interactions From Real Users in the Wild
arXiv cs.AI / 4/23/2026
📰 NewsSignals & Early TrendsModels & Research
Key Points
- The paper introduces SWE-chat, a large-scale “living” dataset of real-world coding agent sessions collected from open-source developers, totaling 6,000 sessions, 63,000+ user prompts, and 355,000+ agent tool calls.
- Analysis of the dataset shows bimodal coding behavior: agents generate virtually all committed code in 41% of sessions (“vibe coding”), while humans author all code in 23% of sessions.
- Even with improving agent capabilities, performance in natural settings is limited: only 44% of agent-produced code makes it into user commits.
- The study finds quality and safety issues, as agent-written code leads to more security vulnerabilities than human-authored code.
- Users frequently resist or correct agent outputs—through corrections, failure reports, and interruptions—in 44% of all interaction turns, motivating a shift from curated benchmarks to evidence-based evaluation.
Related Articles
I’m working on an AGI and human council system that could make the world better and keep checks and balances in place to prevent catastrophes. It could change the world. Really. Im trying to get ahead of the game before an AGI is developed by someone who only has their best interest in mind.
Reddit r/artificial
Deepseek V4 Flash and Non-Flash Out on HuggingFace
Reddit r/LocalLLaMA

DeepSeek V4 Flash & Pro Now out on API
Reddit r/LocalLLaMA

I’m building a post-SaaS app catalog on Base, and here’s what that actually means
Dev.to

r/LocalLLaMa Rule Updates
Reddit r/LocalLLaMA