CAMA: Exploring Collusive Adversarial Attacks in c-MARL

arXiv cs.LG / 2026/3/24

💬 オピニオンSignals & Early TrendsIdeas & Deep AnalysisModels & Research

要点

  • The paper examines new forms of collusive adversarial attacks against cooperative multi-agent reinforcement learning (c-MARL), moving beyond prior work that largely studied single-adversary or white-box perturbation attacks.
  • It proposes three coordinated malicious-agent modes—Collective Malicious Agents, Disguised Malicious Agents, and Spied Malicious Agents—and unifies them under a policy-level framework called CAMA.
  • The authors theoretically analyze attack effectiveness using three criteria: disruptiveness, stealthiness, and attack cost.
  • They implement the attacks via observation information fusion and attack-trigger control, targeting agent policies rather than only isolated observations/actions.
  • Experiments on four SMAC II maps show the attacks can produce additive adversarial synergy, improving outcomes while preserving stealthiness and stability over long horizons.

Abstract

Cooperative multi-agent reinforcement learning (c-MARL) has been widely deployed in real-world applications, such as social robots, embodied intelligence, UAV swarms, etc. Nevertheless, many adversarial attacks still exist to threaten various c-MARL systems. At present, the studies mainly focus on single-adversary perturbation attacks and white-box adversarial attacks that manipulate agents' internal observations or actions. To address these limitations, we in this paper attempt to study collusive adversarial attacks through strategically organizing a set of malicious agents into three collusive attack modes: Collective Malicious Agents, Disguised Malicious Agents, and Spied Malicious Agents. Three novelties are involved: i) three collusive adversarial attacks are creatively proposed for the first time, and a unified framework CAMA for policy-level collusive attacks is designed; ii) the attack effectiveness is theoretically analyzed from the perspectives of disruptiveness, stealthiness, and attack cost; and iii) the three collusive adversarial attacks are technically realized through agent's observation information fusion, attack-trigger control. Finally, multi-facet experiments on four SMAC II maps are performed, and experimental results showcase the three collusive attacks have an additive adversarial synergy, strengthening attack outcome while maintaining high stealthiness and stability over long horizons. Our work fills the gap for collusive adversarial learning in c-MARL.