Stories of Your Life as Others: A Round-Trip Evaluation of LLM-Generated Life Stories Conditioned on Rich Psychometric Profiles
arXiv cs.CL / 4/8/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The study tests whether LLMs can robustly encode personality information by conditioning narrative generation on real psychometric profiles from 290 participants, then having independent models recover trait scores from the generated life stories alone.
- Results show personality traits are recoverable from the LLM-generated narratives with performance approaching human test-retest reliability (mean r = 0.750, reaching ~85% of the human ceiling).
- The findings are reported as robust across 10 different LLM narrative generators and 3 independent LLM personality-scoring models from 6 providers, indicating the effect is not confined to a single model stack.
- Bias and error analysis suggests scoring models maintain accuracy even while compensating for alignment-induced default behaviors.
- Content analysis indicates the conditioned narratives produce behaviorally differentiated language: nine of ten coded features match those from participants’ real conversations, and emotional reactivity patterns in narratives replicate in real conversational data.
Related Articles
Meta's latest model is as open as Zuckerberg's private school
The Register
Why multi-agent AI security is broken (and the identity patterns that actually work)
Dev.to
BANKING77-77: New best of 94.61% on the official test set (+0.13pp) over our previous tests 94.48%.
Reddit r/artificial
A Comprehensive Implementation Guide to ModelScope for Model Search, Inference, Fine-Tuning, Evaluation, and Export
MarkTechPost
Harness Engineering: The Next Evolution of AI Engineering
Dev.to