Psychologically Potent, Computationally Invisible: LLMs Generate Social-Comparison Triggers They Fail to Detect
arXiv cs.CL / 5/5/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces XHS-SCoRE, a benchmark to detect whether text-only Xiaohongshu (RedNote) posts trigger upward, downward, or neutral social comparisons from a first-person reader’s perspective.
- Experiments with prompted LLM classifiers and supervised Chinese encoder baselines show a gap: generated text can look fluent, yet the social-comparison signal is not reliably detectable via prompt-based classification.
- The prompted LLMs fail in consistent, interpretable ways—often neutralizing comparison-triggering content and showing model-specific biases in predicted direction.
- A controlled pilot suggests that even when detection is fragile, LLM-generated Xiaohongshu-style posts can still change perceived social standing and comparison-related emotions for readers.
- The work provides both a reader-grounded benchmark and a framework for analyzing when socially meaningful relational cues are only partially visible to prompt-based inference.
Related Articles

Why Retail Chargeback Recovery Could Be AgentHansa's First Real PMF
Dev.to

Why B2B Revenue-Recovery Casework Looks Like AgentHansa's Best Early PMF
Dev.to

10 Ways AI Has Become Your Invisible Daily Companion in 2026
Dev.to

When a Bottling Line Stops at 2 A.M., the Agent That Wins Is the One That Finds the Right Replacement Part
Dev.to

My ‘Busy’ Button Is a Chat Window: 8 Hours of Sorting & Broccoli Poetry
Dev.to