LLM Agents Predict Social Media Reactions but Do Not Outperform Text Classifiers: Benchmarking Simulation Accuracy Using 120K+ Personas of 1511 Humans
arXiv cs.CL / 4/23/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The study benchmarks LLM-based autonomous agents for predicting specific individuals’ social media reactions (like, dislike, comment, share, no reaction) using 120,000+ persona-content combinations from 1,511 Serbian participants and 27 LLMs.
- In Study 1, the agents reach 70.7% overall accuracy, with LLM choice causing up to a 13 percentage-point spread in performance.
- In Study 2, binary like/dislike evaluation yields MCC=0.29, showing predictive signal beyond chance, but standard TF-IDF text classifiers do better with MCC=0.36.
- The results suggest that any predictive improvement likely comes from semantic access rather than uniquely “agentic” reasoning, while the ability of zero-shot persona-prompted agents raises governance risks of manipulative AI swarms.
- The paper argues these agents could still be useful in simulation for polarization dynamics and policy design, though the findings are limited by single-country sampling and warrant multilingual and fine-tuned follow-ups.
Related Articles
Navigating WooCommerce AI Integrations: Lessons for Agencies & Developers from a Bluehost Conflict
Dev.to
One Day in Shenzhen, Seen Through an AI's Eyes
Dev.to

Underwhelming or underrated? DeepSeek V4 shows “impressive” gains
SCMP Tech

Pics of new rig!
Reddit r/LocalLLaMA
Claude Code: Hooks, Subagents, and Skills — Complete Guide
Dev.to