AI Navigate

From Debate to Deliberation: Structured Collective Reasoning with Typed Epistemic Acts

arXiv cs.AI / 3/13/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Deliberative Collective Intelligence (DCI) for multi-agent LLMs, a phased deliberation approach with typed epistemic acts, a shared workspace, and a convergent flow algorithm that guarantees termination and outputs a structured decision packet containing the selected option, residual objections, minority report, and reopen conditions.
  • DCI defines four reasoning archetypes and 14 typed epistemic acts, enabling differentiated participants to exchange reasoning, preserve disagreements, and produce accountable artifacts.
  • In experiments across 45 tasks and seven domains using Gemini 2.5 Flash, DCI significantly improves non-routine tasks over unstructured debate (+0.95, 95% CI [+0.41, +1.54]) and excels on hidden-profile tasks (9.56, highest of any system on any domain), while underperforming on routine decisions.
  • DCI yields 100% structured decision packets and 98% minority reports, but incurs ~62x single-agent token cost, with single-agent generation still outperforming DCI on overall quality.
  • The study concludes that deliberative structure, not merely more agents, benefits consequential decisions when process accountability justifies the cost.

Abstract

Multi-agent LLM systems increasingly tackle complex reasoning, yet their interaction patterns remain limited to voting, unstructured debate, or pipeline orchestration. None model deliberation: a phased process where differentiated participants exchange typed reasoning moves, preserve disagreements, and converge on accountable outcomes. We introduce Deliberative Collective Intelligence (DCI), specifying four reasoning archetypes, 14 typed epistemic acts, a shared workspace, and DCI-CF, a convergent flow algorithm that guarantees termination with a structured decision packet containing the selected option, residual objections, minority report, and reopen conditions. We evaluate on 45 tasks across seven domains using Gemini 2.5 Flash. On non-routine tasks (n=40), DCI significantly improves over unstructured debate (+0.95, 95% CI [+0.41, +1.54]). DCI excels on hidden-profile tasks requiring perspective integration (9.56, highest of any system on any domain) while failing on routine decisions (5.39), confirming task-dependence. DCI produces 100% structured decision packets and 98% minority reports, artifacts absent from all baselines. However, DCI consumes ~62x single-agent tokens, and single-agent generation outperforms DCI on overall quality. DCI's contribution is not that more agents are better, but that consequential decisions benefit from deliberative structure when process accountability justifies the cost.