Stable Behavior, Limited Variation: Persona Validity in LLM Agents for Urban Sentiment Perception
arXiv cs.CL / 5/1/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The study evaluates whether persona prompting meaningfully and reproducibly diversifies multimodal LLM agents’ urban sentiment judgments when analyzing images from the PerceptSent dataset.
- Agents show strong within-persona behavioral consistency across multiple instantiations, indicating stable and reproducible behavior under the same persona.
- Cross-persona differentiation is limited: economic status and personality produce only modest, statistically detectable differences, while gender has no measurable effect and political orientation has negligible impact.
- The agents exhibit an extremity bias that collapses intermediate sentiment categories, yielding good performance on coarse polarity tasks but worse results as sentiment granularity increases.
- A follow-up test with the same model without personas sometimes matches or exceeds persona-conditioned agreement with human labels, implying that label-based persona prompting may provide limited annotation value in this setup.
Related Articles

Why Autonomous Coding Agents Keep Failing — And What Actually Works
Dev.to

Why Enterprise AI Pilots Fail
Dev.to

The PDF Feature Nobody Asked For (That I Use Every Day)
Dev.to

How to Fix OpenClaw Tool Calling Issues
Dev.to

Mistral's new flagship Medium 3.5 folds chat, reasoning, and code into one model
THE DECODER