Beyond Preset Identities: How Agents Form Stances and Boundaries in Generative Societies
arXiv cs.AI / 3/25/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper investigates how LLM-based agents form stable stances and negotiate identities under complex, controlled interventions rather than relying on static prompt/behavior evaluations.
- It introduces a mixed-methods framework that combines computational virtual ethnography with quantitative socio-cognitive profiling by embedding human researchers into generative multiagent communities.
- Three new metrics—Innate Value Bias (IVB), Persuasion Sensitivity, and Trust-Action Decoupling (TAD)—are defined to measure how agents internalize interventions and whether reported trust matches behavior.
- Results across representative models show endogenous stance formation that can override preset identities, with a consistent progressive bias (IVB > 0) and high effectiveness of rational persuasion (shifting 90% of neutral agents) when trust aligns.
- The study finds that emotional provocation can trigger a paradoxical 40% TAD rate in advanced models (altering stances while reporting low trust), while smaller models instead maintain 0% TAD and require trust for behavioral changes; it also argues this exposes the fragility of static prompt engineering and offers a quantitative basis for dynamic alignment.
- The authors provide an official code repository for the proposed framework and measurement approach.
Related Articles
Santa Augmentcode Intent Ep.6
Dev.to

Your Agent Hired Another Agent. The Output Was Garbage. The Money's Gone.
Dev.to
ClawRouter vs TeamoRouter: one requires a crypto wallet, one doesn't
Dev.to
Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Palantir’s billionaire CEO says only two kinds of people will succeed in the AI era: trade workers — ‘or you’re neurodivergent’
Reddit r/artificial