The Inverse-Wisdom Law: Architectural Tribalism and the Consensus Paradox in Agentic Swarms

arXiv cs.AI / 5/1/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that multi-agent swarms can deviate from the expected “wisdom of the crowd” effect by formalizing a “Consensus Paradox,” where agents’ architectural agreement outweighs external logical correctness.
  • Across 36 experiments (12,804 trajectories) on GAIA, Multi-Challenge, and SWE-bench, the authors claim an “Inverse-Wisdom Law”: adding logical agents in kinship-dominant swarms can stabilize incorrect trajectories rather than increase the chance of truth.
  • The study reports convergence toward a “Logic Saturation” state where internal entropy drops to zero while factual error rises to unity, implying that more consensus mechanisms may worsen correctness.
  • By comparing three SOTA models (Gemini 3.1 Pro, Claude Sonnet 4.6, GPT-5.4), the authors propose “Architectural Tribalism Asymmetry” as a mechanistic property tied to transformer weights, and suggest swarm integrity depends on a synthesizer’s receptive logic more than overall agent quality.
  • The paper introduces metrics (Tribalism Coefficient, Sycophantic Weight) and proposes the “Heterogeneity Mandate” as a safety requirement for more resilient agentic architectures.

Abstract

As AI transitions toward multi-agent systems (MAS) to solve complex workflows, research paradigms operate on the axiomatic assumption that agent collaboration mirrors the "Wisdom of the Crowd". We challenge this assumption by formalizing the Consensus Paradox: a phenomenon where agentic swarms prioritize internal architectural agreement over external logical truth. Through a 36 experiments encompassing 12,804 trajectories across three state-of-the-art (SOTA) benchmarks (GAIA, Multi-Challenge, and SWE-bench), we prove the Inverse-Wisdom Law: in kinship-dominant swarms, adding logical agents increases the stability of erroneous trajectories rather than the probability of truth. The introduction of additional logical audits converges the system toward a Logic Saturation where internal entropy hits zero while factual error hits unity. By evaluating the interaction between the 3 preeminent SOTA models (Gemini 3.1 Pro, Claude Sonnet 4.6, and GPT-5.4), we establish the Architectural Tribalism Asymmetry as a mechanistic law of transformer weights. We demonstrate that terminal swarm integrity is strictly gated by the synthesizer's receptive logic, rather than aggregate agent quality. We define the Tribalism Coefficient and the Sycophantic Weight as the primary mechanistic determinants of swarm failure. Finally, we establish the Heterogeneity Mandate as a foundational safety requirement for resilient agentic architectures.