CAMA: Exploring Collusive Adversarial Attacks in c-MARL
arXiv cs.LG / 3/24/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper examines new forms of collusive adversarial attacks against cooperative multi-agent reinforcement learning (c-MARL), moving beyond prior work that largely studied single-adversary or white-box perturbation attacks.
- It proposes three coordinated malicious-agent modes—Collective Malicious Agents, Disguised Malicious Agents, and Spied Malicious Agents—and unifies them under a policy-level framework called CAMA.
- The authors theoretically analyze attack effectiveness using three criteria: disruptiveness, stealthiness, and attack cost.
- They implement the attacks via observation information fusion and attack-trigger control, targeting agent policies rather than only isolated observations/actions.
- Experiments on four SMAC II maps show the attacks can produce additive adversarial synergy, improving outcomes while preserving stealthiness and stability over long horizons.
Related Articles
How AI is Transforming Dynamics 365 Business Central
Dev.to
Algorithmic Gaslighting: A Formal Legal Template to Fight AI Safety Pivots That Cause Psychological Harm
Reddit r/artificial
Do I need different approaches for different types of business information errors?
Dev.to
ShieldCortex: What We Learned Protecting AI Agent Memory
Dev.to
WordPress Theme Customization Without Code: The AI Revolution
Dev.to