Learning-guided Prioritized Planning for Lifelong Multi-Agent Path Finding in Warehouse Automation

arXiv cs.RO / 3/26/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper targets Lifelong Multi-Agent Path Finding (MAPF) for warehouse automation, aiming to maintain conflict-free multi-robot routing over long-term, dynamic conditions to improve overall throughput.
  • It proposes RL-RH-PP, a framework that combines reinforcement learning with classical Rolling Horizon Prioritized Planning by using learning-based priority assignment over time rather than replacing search entirely.
  • The method formulates dynamic priority assignment as a Partially Observable Markov Decision Process (POMDP) and uses an attention-based autoregressive neural network to decode priority orders on-the-fly for sequential single-agent planning.
  • Experiments in realistic warehouse simulations reportedly show RL-RH-PP achieves the highest total throughput versus baselines and generalizes across different agent densities, planning horizons, and warehouse layouts.
  • Interpretive analyses suggest the learned priority policy proactively manages congestion by prioritizing congested agents and redirecting flow to ease traffic patterns.

Abstract

Lifelong Multi-Agent Path Finding (MAPF) is critical for modern warehouse automation, which requires multiple robots to continuously navigate conflict-free paths to optimize the overall system throughput. However, the complexity of warehouse environments and the long-term dynamics of lifelong MAPF often demand costly adaptations to classical search-based solvers. While machine learning methods have been explored, their superiority over search-based methods remains inconclusive. In this paper, we introduce Reinforcement Learning (RL) guided Rolling Horizon Prioritized Planning (RL-RH-PP), the first framework integrating RL with search-based planning for lifelong MAPF. Specifically, we leverage classical Prioritized Planning (PP) as a backbone for its simplicity and flexibility in integrating with a learning-based priority assignment policy. By formulating dynamic priority assignment as a Partially Observable Markov Decision Process (POMDP), RL-RH-PP exploits the sequential decision-making nature of lifelong planning while delegating complex spatial-temporal interactions among agents to reinforcement learning. An attention-based neural network autoregressively decodes priority orders on-the-fly, enabling efficient sequential single-agent planning by the PP planner. Evaluations in realistic warehouse simulations show that RL-RH-PP achieves the highest total throughput among baselines and generalizes effectively across agent densities, planning horizons, and warehouse layouts. Our interpretive analyses reveal that RL-RH-PP proactively prioritizes congested agents and strategically redirects agents from congestion, easing traffic flow and boosting throughput. These findings highlight the potential of learning-guided approaches to augment traditional heuristics in modern warehouse automation.