MARL-GPT: Foundation Model for Multi-Agent Reinforcement Learning

arXiv cs.AI / 4/8/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces MARL-GPT, a GPT-based foundation model designed to learn and perform across multiple multi-agent reinforcement learning (MARL) environments and tasks using a single model rather than task-specific architectures.
  • MARL-GPT is trained via offline reinforcement learning on large-scale expert trajectories (400M for SMACv2, 100M for GRF, and 1B for POGEMA) and uses a single transformer-based observation encoder that avoids task-specific tuning.
  • Experiments indicate that MARL-GPT delivers competitive results against specialized MARL baselines across the tested benchmarks, including StarCraft Multi-Agent Challenge, Google Research Football, and POGEMA.
  • The authors argue the approach supports the broader goal of a “foundation” MARL model that generalizes across significantly different multi-agent problem settings, analogous to how LLMs generalize across NLP tasks.

Abstract

Recent advances in multi-agent reinforcement learning (MARL) have demonstrated success in numerous challenging domains and environments, but typically require specialized models for each task. In this work, we propose a coherent methodology that makes it possible for a single GPT-based model to learn and perform well across diverse MARL environments and tasks, including StarCraft Multi-Agent Challenge, Google Research Football and POGEMA. Our method, MARL-GPT, applies offline reinforcement learning to train at scale on the expert trajectories (400M for SMACv2, 100M for GRF, and 1B for POGEMA) combined with a single transformer-based observation encoder that requires no task-specific tuning. Experiments show that MARL-GPT achieves competitive performance compared to specialized baselines in all tested environments. Thus, our findings suggest that it is, indeed, possible to build a multi-task transformer-based model for a wide variety of (significantly different) multi-agent problems paving the way to the fundamental MARL model (akin to ChatGPT, Llama, Mistral etc. in natural language modeling).