DeliberationBench: A Normative Benchmark for the Influence of Large Language Models on Users' Views
arXiv cs.AI / 3/12/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- DeliberationBench is proposed as a normative benchmark for assessing the persuasive influence of large language models (LLMs) on users' beliefs, using deliberative opinion polling as the standard.
- The authors demonstrate the approach with a preregistered randomized experiment involving 4,088 U.S. participants who discussed 65 policy proposals with six frontier LLMs.
- Results indicate substantial influence from the tested LLMs on participants' opinions, and this influence is positively associated with net opinion shifts after deliberation, suggesting broadly epistemically desirable effects.
- The analysis finds differential influence across topic areas, demographic subgroups, and model variants, highlighting nuanced patterns in how LLMs shape viewpoints.
- The framework is presented as an evaluation and monitoring tool to ensure LLM influence remains aligned with democratically legitimate standards and preserves users’ autonomy in forming their views.
Related Articles
How CVE-2026-25253 exposed every OpenClaw user to RCE — and how to fix it in one command
Dev.to
Does Synthetic Data Generation of LLMs Help Clinical Text Mining?
Dev.to
What CVE-2026-25253 Taught Me About Building Safe AI Assistants
Dev.to
Day 52: Building vs Shipping — Why We Had 711 Commits and 0 Users
Dev.to
The Dawn of the Local AI Era: From iPhone 17 Pro to the Future of NVIDIA RTX
Dev.to