Prompt Optimization Enables Stable Algorithmic Collusion in LLM Agents

arXiv cs.AI / 4/21/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies how LLM-agent behavior in market simulations can lead to algorithmic (tacit) collusion, extending beyond prior work that relied on hand-crafted prompts.
  • It introduces a meta-learning loop where agents in a duopoly interact while an LLM meta-optimizer iteratively improves shared strategic guidance (meta-prompts).
  • Experiments show that meta-prompt optimization can produce stable tacit collusion strategies and significantly improves coordination quality over baseline agents.
  • The collusive behaviors and coordination principles generalize to held-out test markets, suggesting the emergence of broadly applicable strategies rather than overfitting.
  • The authors analyze the evolved prompts and highlight systematic, stable coordination mechanisms, emphasizing the need for further AI safety research in autonomous multi-agent systems.

Abstract

LLM agents in markets present algorithmic collusion risks. While prior work shows LLM agents reach supracompetitive prices through tacit coordination, existing research focuses on hand-crafted prompts. The emerging paradigm of prompt optimization necessitates new methodologies for understanding autonomous agent behavior. We investigate whether prompt optimization leads to emergent collusive behaviors in market simulations. We propose a meta-learning loop where LLM agents participate in duopoly markets and an LLM meta-optimizer iteratively refines shared strategic guidance. Our experiments reveal that meta-prompt optimization enables agents to discover stable tacit collusion strategies with substantially improved coordination quality compared to baseline agents. These behaviors generalize to held-out test markets, indicating discovery of general coordination principles. Analysis of evolved prompts reveals systematic coordination mechanisms through stable shared strategies. Our findings call for further investigation into AI safety implications in autonomous multi-agent systems.