Cooperate to Compete: Strategic Coordination in Multi-Agent Conquest

arXiv cs.CL / 4/29/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces “Cooperate to Compete (C2C),” a multi-agent game where LM-based agents must mix short-term private cooperation with long-term competitive objectives under asymmetric, secret goals.
  • C2C allows non-binding negotiations, so alliances can form and dissolve as players’ incentives shift, creating a realistic mixed-motive coordination setting.
  • The authors compare human vs. LM negotiation behavior, finding humans prefer lower-complexity deals, are less reliable partners, and accept proposals without counteroffers less often (56.3% vs. 67.6%).
  • By using prompting informed by these behavioral differences to adjust agent negotiation strategies, the study improves win rates from 22.2% to 32.7%.
  • With more than 1,100 games, 16,000+ private conversations, and 15.2M tokens, the authors position C2C as a testbed for building and evaluating LM agents for real-world deployment coordination.

Abstract

Language Model (LM)-based agents remain largely untested in mixed-motive settings where agents must leverage short-term cooperation for long-term competitive goals (e.g., multi-party politics). We introduce Cooperate to Compete (C2C), a multi-agent environment where players can engage in private negotiations while competing to be the first to achieve their secret objective. Players have asymmetric objectives and negotiations are non-binding, allowing alliances to form and break as players' short-term interests align and diverge. We run AI only games and conduct a user study pitting human players against AI opponents. We identify significant differences between human and AI negotiation behaviors, finding that humans favor lower-complexity deals and are significantly less reliable partners compared to LM-based agents. We also find that humans are more aggressive negotiators, accepting deals without a counteroffer only 56.3% of the time compared to 67.6% for LM-based agents. Through targeted prompting inspired by these findings, we modify agents' negotiation behavior and improve win rates from 22.2% to 32.7%. We run over 1,100 games with over 16,000 private conversations totaling 15.2 million tokens and over 150,000 player actions. Our results establish C2C as a testbed for studying and building LM-based agents that can navigate the sophisticated coordination required for real-world deployments. The game, code, and dataset may be found at https://negotiationgame.io/c2c.