Polarization by Default: Auditing Recommendation Bias in LLM-Based Content Curation
arXiv cs.CL / 4/20/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The study audits how large language models (LLMs) introduce and structure bias when curation and ranking are applied to real social-media posts from Twitter/X, Bluesky, and Reddit.
- Across 540,000 simulated top-10 selections using six prompting strategies and three major providers (OpenAI, Anthropic, Google), the authors find that some biases are robust while others are highly sensitive to prompt design.
- Results show amplified polarization in all configurations, a strong prompt-dependent inversion in toxicity handling between engagement-focused and information-focused prompts, and predominantly negative sentiment bias.
- Provider-level comparisons reveal distinct trade-offs: GPT-4o Mini is most consistent across prompts, Claude and Gemini adapt more strongly for toxicity handling, and Gemini most strongly prefers negative sentiment.
- On Twitter/X, political-leaning bias is the clearest demographic signal: left-leaning authors are systematically over-represented in selections even when right-leaning authors make up the largest share of the candidate pool, and this persists across prompts.
Related Articles
Which Version of Qwen 3.6 for M5 Pro 24g
Reddit r/LocalLLaMA

From Theory to Reality: Why Most AI Agent Projects Fail (And How Mine Did Too)
Dev.to

GPT-5.4-Cyber: OpenAI's Game-Changer for AI Security and Defensive AI
Dev.to

Building Digital Souls: The Brutal Reality of Creating AI That Understands You Like Nobody Else
Dev.to
Local LLM Beginner’s Guide (Mac - Apple Silicon)
Reddit r/artificial