KD-MARL: Resource-Aware Knowledge Distillation in Multi-Agent Reinforcement Learning

arXiv cs.AI / 4/10/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces KD-MARL, a two-stage resource-aware knowledge distillation framework to deploy multi-agent reinforcement learning (MARL) on edge or embedded platforms with strict compute, memory, and inference-time limits.
  • KD-MARL distills both action-level behavior and coordination structure from centralized expert policies to lightweight decentralized student agents, enabling training without a critic via distilled advantage signals and structured policy supervision.
  • The method is designed to handle heterogeneous agents, letting each student model scale its capacity to match its observation complexity under partial or limited observability.
  • Experiments on SMAC and MPE benchmarks show strong performance retention, preserving over 90% of expert performance while cutting computational cost by up to 28.6× FLOPs.
  • Overall, the work claims expert-level coordination can be maintained through structured distillation while enabling practical, cost-efficient MARL execution in resource-constrained settings.

Abstract

Real world deployment of multi agent reinforcement learning MARL systems is fundamentally constrained by limited compute memory and inference time. While expert policies achieve high performance they rely on costly decision cycles and large scale models that are impractical for edge devices or embedded platforms. Knowledge distillation KD offers a promising path toward resource aware execution but existing KD methods in MARL focus narrowly on action imitation often neglecting coordination structure and assuming uniform agent capabilities. We propose resource aware Knowledge Distillation for Multi Agent Reinforcement Learning KD MARL a two stage framework that transfers coordinated behavior from a centralized expert to lightweight decentralized student agents. The student policies are trained without a critic relying instead on distilled advantage signals and structured policy supervision to preserve coordination under heterogeneous and limited observations. Our approach transfers both action level behavior and structural coordination patterns from expert policies while supporting heterogeneous student architectures allowing each agent model capacity to match its observation complexity which is crucial for efficient execution under partial or limited observability and limited onboard resources. Extensive experiments on SMAC and MPE benchmarks demonstrate that KD MARL achieves high performance retention while substantially reducing computational cost. Across standard multi agent benchmarks KD MARL retains over 90 percent of expert performance while reducing computational cost by up to 28.6 times FLOPs. The proposed approach achieves expert level coordination and preserves it through structured distillation enabling practical MARL deployment across resource constrained onboard platforms.