DeliberationBench: A Normative Benchmark for the Influence of Large Language Models on Users' Views
arXiv cs.AI / 3/12/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- DeliberationBench is proposed as a normative benchmark for assessing the persuasive influence of large language models (LLMs) on users' beliefs, using deliberative opinion polling as the standard.
- The authors demonstrate the approach with a preregistered randomized experiment involving 4,088 U.S. participants who discussed 65 policy proposals with six frontier LLMs.
- Results indicate substantial influence from the tested LLMs on participants' opinions, and this influence is positively associated with net opinion shifts after deliberation, suggesting broadly epistemically desirable effects.
- The analysis finds differential influence across topic areas, demographic subgroups, and model variants, highlighting nuanced patterns in how LLMs shape viewpoints.
- The framework is presented as an evaluation and monitoring tool to ensure LLM influence remains aligned with democratically legitimate standards and preserves users’ autonomy in forming their views.
Related Articles

Hey dev.to community – sharing my journey with Prompt Builder, Insta Posts, and practical SEO
Dev.to

How to Build Passive Income with AI in 2026: A Developer's Practical Guide
Dev.to

The Research That Doesn't Exist
Dev.to

Jeff Bezos reportedly wants $100 billion to buy and transform old manufacturing firms with AI
TechCrunch

Krish Naik: AI Learning Path For 2026- Data Science, Generative and Agentic AI Roadmap
Dev.to