The Persuasion Paradox: When LLM Explanations Fail to Improve Human-AI Team Performance
arXiv cs.AI / 4/7/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces a “Persuasion Paradox,” arguing that LLM explanations can increase users’ confidence and reliance without reliably improving—sometimes even reducing—task accuracy in human-AI teams.
- Across three controlled studies (RAVEN visual reasoning and LSAT-style logical reasoning), explanation-based interfaces improved confidence but often failed to beat accuracy obtained from AI predictions alone, and they weakened users’ ability to correct model errors.
- For visual reasoning, interfaces that show model uncertainty (e.g., predicted probabilities) and use selective automation to defer uncertain cases to humans achieved higher accuracy and better error recovery than explanations.
- For language-based logical reasoning, however, LLM explanations produced the best accuracy and recovery, outperforming both probability-based support and expert-written explanations, indicating strong task-dependent effects.
- The authors conclude that subjective measures like trust and perceived clarity are poor proxies for performance and recommend designing interaction systems that emphasize calibrated reliance and error recovery rather than persuasive fluency.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.
Related Articles

Black Hat Asia
AI Business
v0.20.5
Ollama Releases

Inside Anthropic's Project Glasswing: The AI Model That Found Zero-Days in Every Major OS
Dev.to
Gemma 4 26B fabricated an entire code audit. I have the forensic evidence from the database.
Reddit r/LocalLLaMA
SoloEngine: Low-Code Agentic AI Development Platform with Native Support for Multi-Agent Collaboration, MCP, and Skill System
Dev.to