AI Navigate

Scale-Plan: Scalable Language-Enabled Task Planning for Heterogeneous Multi-Robot Teams

arXiv cs.AI / 3/11/2026

Ideas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • Scale-Plan is a scalable framework that leverages large language models (LLMs) to assist in task planning for heterogeneous multi-robot teams by generating compact, task-relevant problem representations from natural language instructions.
  • It combines a PDDL domain specification with shallow LLM reasoning and structured graph search to filter irrelevant perceptions, improving the efficiency of long-horizon task planning in object-rich environments.
  • Scale-Plan addresses limitations of traditional symbolic planners and pure LLM approaches by enhancing grounding and reducing hallucinations, thus enabling better plan generation, decomposition, and allocation.
  • The approach was evaluated on complex multi-agent tasks using a new benchmark, MAT2-THOR, demonstrating superior scalability and reliability compared to baseline methods including pure LLM and hybrid LLM-PDDL planners.
  • This work advances multi-robot collaborative planning by improving scalability and adaptability in environments with abundant perceptual data, enabling more robust real-world deployment of heterogeneous robotic teams.

Computer Science > Robotics

arXiv:2603.08814 (cs)
[Submitted on 9 Mar 2026]

Title:Scale-Plan: Scalable Language-Enabled Task Planning for Heterogeneous Multi-Robot Teams

View a PDF of the paper titled Scale-Plan: Scalable Language-Enabled Task Planning for Heterogeneous Multi-Robot Teams, by Piyush Gupta and 3 other authors
View PDF HTML (experimental)
Abstract:Long-horizon task planning for heterogeneous multi-robot systems is essential for deploying collaborative teams in real-world environments; yet, it remains challenging due to the large volume of perceptual information, much of which is irrelevant to task objectives and burdens planning. Traditional symbolic planners rely on manually constructed problem specifications, limiting scalability and adaptability, while recent large language model (LLM)-based approaches often suffer from hallucinations and weak grounding-i.e., poor alignment between generated plans and actual environmental objects and constraints-in object-rich settings. We present Scale-Plan, a scalable LLM-assisted framework that generates compact, task-relevant problem representations from natural language instructions. Given a PDDL domain specification, Scale-Plan constructs an action graph capturing domain structure and uses shallow LLM reasoning to guide a structured graph search that identifies a minimal subset of relevant actions and objects. By filtering irrelevant information prior to planning, Scale-Plan enables efficient decomposition, allocation, and long-horizon plan generation. We evaluate our approach on complex multi-agent tasks and introduce MAT2-THOR, a cleaned benchmark built on AI2-THOR for reliable evaluation of multi-robot planning systems. Scale-Plan outperforms pure LLM and hybrid LLM-PDDL baselines across all metrics, improving scalability and reliability.
Subjects: Robotics (cs.RO); Artificial Intelligence (cs.AI); Emerging Technologies (cs.ET); Multiagent Systems (cs.MA)
Cite as: arXiv:2603.08814 [cs.RO]
  (or arXiv:2603.08814v1 [cs.RO] for this version)
  https://doi.org/10.48550/arXiv.2603.08814
Focus to learn more
arXiv-issued DOI via DataCite

Submission history

From: Piyush Gupta [view email]
[v1] Mon, 9 Mar 2026 18:13:18 UTC (2,359 KB)
Full-text links:

Access Paper:

    View a PDF of the paper titled Scale-Plan: Scalable Language-Enabled Task Planning for Heterogeneous Multi-Robot Teams, by Piyush Gupta and 3 other authors
  • View PDF
  • HTML (experimental)
  • TeX Source
Current browse context:
cs.RO
< prev   |   next >
Change to browse by:

References & Citations

export BibTeX citation Loading...

BibTeX formatted citation

×
Data provided by:

Bookmark

BibSonomy logo Reddit logo
Bibliographic Tools

Bibliographic and Citation Tools

Bibliographic Explorer Toggle
Bibliographic Explorer (What is the Explorer?)
Connected Papers Toggle
Connected Papers (What is Connected Papers?)
Litmaps Toggle
Litmaps (What is Litmaps?)
scite.ai Toggle
scite Smart Citations (What are Smart Citations?)
Code, Data, Media

Code, Data and Media Associated with this Article

alphaXiv Toggle
alphaXiv (What is alphaXiv?)
Links to Code Toggle
CatalyzeX Code Finder for Papers (What is CatalyzeX?)
DagsHub Toggle
DagsHub (What is DagsHub?)
GotitPub Toggle
Gotit.pub (What is GotitPub?)
Huggingface Toggle
Hugging Face (What is Huggingface?)
Links to Code Toggle
Papers with Code (What is Papers with Code?)
ScienceCast Toggle
ScienceCast (What is ScienceCast?)
Demos

Demos

Replicate Toggle
Replicate (What is Replicate?)
Spaces Toggle
Hugging Face Spaces (What is Spaces?)
Spaces Toggle
TXYZ.AI (What is TXYZ.AI?)
Related Papers

Recommenders and Search Tools

Link to Influence Flower
Influence Flower (What are Influence Flowers?)
Core recommender toggle
CORE Recommender (What is CORE?)
About arXivLabs

arXivLabs: experimental projects with community collaborators

arXivLabs is a framework that allows collaborators to develop and share new arXiv features directly on our website.

Both individuals and organizations that work with arXivLabs have embraced and accepted our values of openness, community, excellence, and user data privacy. arXiv is committed to these values and only works with partners that adhere to them.

Have an idea for a project that will add value for arXiv's community? Learn more about arXivLabs.