MARLIN: Multi-Agent Reinforcement Learning Guided by Language-Based Inter-Robot Negotiation

arXiv cs.RO / 4/14/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • MARLINは、多機械(マルチロボット)向けの強化学習を、言語によるロボット間交渉(LLM)で安全性と探索効率を高めるハイブリッド枠組みとして提案している。
  • 具体的には、強化学習ポリシーが十分学習する前にLLMが高レベル計画を行い、言語で交渉・計画を生成してポリシー学習を導く。
  • 学習中は強化学習とLLMベースの交渉(計画)を動的に切り替え、初期段階の危険な挙動につながりうる探索を抑える設計になっている。
  • シミュレーションだけでなく実機ロボットでも評価し、ローカル/リモート双方の言語モデルを用いて、標準的なマルチエージェント強化学習より初期性能を高めつつ最終性能を低下させない結果を報告している。

Abstract

Multi-agent reinforcement learning is a key method for training multi-robot systems. Through rewarding or punishing robots over a series of episodes according to their performance, they can be trained and then deployed in the real world. However, poorly trained policies can lead to unsafe behaviour during early training stages. We introduce Multi-Agent Reinforcement Learning guided by language-based Inter-robot Negotiation (MARLIN), a hybrid framework in which large language models provide high-level planning before the reinforcement learning policy has learned effective behaviours. Robots use language models to negotiate actions and generate plans that guide policy learning. The system dynamically switches between reinforcement learning and language-model-based negotiation during training, enabling safer and more effective exploration. MARLIN is evaluated using both simulated and physical robots with local and remote language models. Results show that, compared to standard multi-agent reinforcement learning, the hybrid approach achieves higher performance in early training without reducing final performance. The code is available at https://github.com/SooratiLab/MARLIN.