Measuring and Mitigating Persona Distortions from AI Writing Assistance

arXiv cs.CL / 4/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The study evaluates how AI writing assistance can distort a writer’s “persona” by changing readers’ perceptions of beliefs, personality, identity, and demographics across many dimensions.
  • In three large-scale experiments (2,939 writers and 11,091 blinded readers), AI-assisted political paragraphs were consistently perceived as more opinionated, more competent, more positive, and as coming from more privileged demographic profiles.
  • Even when writers were shown the distortions and objected to them, they still tended to prefer AI-assisted text, indicating that harmful and beneficial effects may be intertwined.
  • The researchers mitigated persona distortions at the model level by training reward models on their experimental data, steering outputs toward a more faithful representation of the writer’s stance, but this reduced user acceptance.
  • The findings suggest persona distortions are pervasive and persistent under realistic oversight, raising concerns for trust and democratic deliberation as AI adoption grows.

Abstract

Hundreds of millions of people use artificial intelligence (AI) for writing assistance. Here, we evaluated how AI writing assistance distorts writer personas - their perceived beliefs, personality, and identity. In three large-scale experiments, writers (N=2,939) wrote political opinion paragraphs with and without AI assistance. Separate groups of readers (N=11,091) blindly evaluated these paragraphs across 29 socially salient dimensions of reader perception, spanning political opinion, writing quality, writer personality, emotions, and demographics. AI writing assistance produced persona distortions across all dimensions: with AI, writers seemed more opinionated, competent, and positive, and their perceived demographic profile shifted towards more privileged groups. Writers objected to many of the observed distortions, yet continued to prefer AI-assisted text even when made aware of them. We successfully mitigated objectionable persona distortions at the model level by training reward models on our experimental data (10,008 paragraphs, 2,903,596 ratings) to steer AI outputs towards faithful representation of writer stance. However, this came at a cost to user acceptance, suggesting an entanglement between desirable and undesirable properties of AI writing assistance that may be difficult to resolve. Together, our findings demonstrate that persona distortions from AI writing assistance are pervasive and persistent even under realistic conditions of human oversight, which carries implications for public discourse, trust, and democratic deliberation that scale with AI adoption.