Large Language Models for Multi-Robot Systems: A Survey

arXiv cs.RO / 5/5/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The survey reviews how large language models (LLMs) can be integrated into multi-robot systems (MRS) to improve communication, task allocation, planning, and human-robot interaction.
  • It organizes LLM-enabled approaches by level—high-level task allocation, mid-level motion planning, low-level action generation, and human intervention—showing their coverage across multiple robotics application domains.
  • The paper highlights key application areas such as household robotics, construction, formation control, target tracking, and robot games to demonstrate the breadth of LLM use cases in MRS.
  • It discusses major limitations that hinder real-world MRS adoption, including weak mathematical reasoning, hallucinations, latency constraints, and the need for robust benchmarking.
  • The authors outline future research directions, focusing on improved fine-tuning, reasoning techniques, and task-specific models, and they keep the associated paper list updated via an open-source GitHub repository.

Abstract

The rapid advancement of Large Language Models (LLMs) has opened new possibilities in Multi-Robot Systems (MRS), enabling enhanced communication, task allocation and planning, and human-robot interaction. Unlike traditional single-robot and multi-agent systems, MRS poses unique challenges, including coordination, scalability, and real-world adaptability. This survey provides the first dedicated review of LLM integration into MRS. It systematically categorizes their applications across high-level task allocation, mid-level motion planning, low-level action generation, and human intervention. We highlight key applications in diverse domains, such as household robotics, construction, formation control, target tracking, and robot games, showcasing the versatility and transformative potential of LLMs in MRS. Furthermore, we examine the challenges that limit adapting LLMs to MRS, including mathematical reasoning limitations, hallucination, latency issues, and the need for robust benchmarking systems. Finally, we outline opportunities for future research, emphasizing advancements in fine-tuning, reasoning techniques, and task-specific models. This survey aims to guide researchers in the intelligence and real-world deployment of MRS powered by LLMs. Given the rapidly evolving nature of research in the field, we continuously update the paper list in the open-source GitHub repository.