The Price of Paranoia: Robust Risk-Sensitive Cooperation in Non-Stationary Multi-Agent Reinforcement Learning
arXiv cs.AI / 4/20/2026
💬 OpinionModels & Research
Key Points
- Cooperative equilibria in multi-agent reinforcement learning can become unstable because agents co-learn: each agent’s gradient updates change its partner’s action distribution, adding noise exactly where cooperation is most sensitive.
- The paper shows that even strongly Pareto-dominant cooperative equilibria are exponentially unstable under standard risk-neutral learning, and collapse irreversibly once partner-induced noise exceeds a critical threshold.
- Applying “distributional robustness” in a naive way (e.g., making agents risk-averse over return distributions) can worsen instability because it penalizes high-variance cooperative actions more than defection.
- The authors propose robustness aimed at the policy-gradient update variance caused by partner uncertainty, using an online measure of partner unpredictability to modulate gradients and expand the cooperation basin.
- They introduce metrics—“Price of Paranoia” and a “Cooperation Window”—to jointly characterize stability, sample efficiency, and welfare recovery, and derive an optimal robustness level as a closed-form trade-off.
Related Articles

From Theory to Reality: Why Most AI Agent Projects Fail (And How Mine Did Too)
Dev.to

GPT-5.4-Cyber: OpenAI's Game-Changer for AI Security and Defensive AI
Dev.to
Local LLM Beginner’s Guide (Mac - Apple Silicon)
Reddit r/artificial

Is Your Skill Actually Good? Systematically Validating Agent Skills with Evals
Dev.to

Space now with memory
Dev.to