Automated Adversarial Collaboration for Advancing Theory Building in the Cognitive Sciences

arXiv cs.AI / 4/29/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes an automated adversarial collaboration framework to arbitrate between competing cognitive science theories while also discovering candidate models and experiments during the evaluation process.
  • The approach uses an LLM-based “theory agent” system combined with program synthesis and information-theoretic experimental design inside a closed-loop workflow.
  • In simulations across three classic categorization theories, the method successfully recovered the ground-truth theory under various noise conditions, though performance weakened in the hardest cases.
  • The authors present the framework as a concrete proof of concept for closed-loop, in-silico theory adjudication in cognitive science, addressing limitations of narrow paradigms and local comparisons.

Abstract

Cognitive science often evaluates theories through narrow paradigms and local model comparisons, limiting the integration of evidence across tasks and realizations. We introduce an automated adversarial collaboration framework for adjudicating among competing theories even when the candidate models and experiments must be discovered during the adjudication process. The system combines LLM-based theory agents, program synthesis, and information-theoretic experimental design in a closed loop. In a simulation study spanning three classic categorization theories, the framework recovered the ground-truth theory across noise settings with weaker reliability in the hardest settings. Together, the framework and findings provide a concrete proof of concept for closed-loop, in-silico theory adjudication in cognitive science.