When Should a Robot Think? Resource-Aware Reasoning via Reinforcement Learning for Embodied Robotic Decision-Making

arXiv cs.RO / 3/27/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses a core problem for embodied robotic agents using LLM-style reasoning: reasoning can improve decisions but also adds latency and resource overhead that can harm reliability and task success.
  • It introduces RARRL (Resource-Aware Reasoning via Reinforcement Learning), a hierarchical framework that learns a high-level orchestration policy to decide when to invoke reasoning, which reasoning module to use, and how much compute budget to allocate.
  • The orchestration policy adapts based on current observations, execution history, and remaining resources rather than relying on fixed schedules or hand-crafted heuristics.
  • Experiments using empirical latency profiles derived from the ALFRED benchmark show RARRL improves task success rates while reducing execution latency and increasing robustness compared with baseline reasoning strategies.

Abstract

Embodied robotic systems increasingly rely on large language model (LLM)-based agents to support high-level reasoning, planning, and decision-making during interactions with the environment. However, invoking LLM reasoning introduces substantial computational latency and resource overhead, which can interrupt action execution and reduce system reliability. Excessive reasoning may delay actions, while insufficient reasoning often leads to incorrect decisions and task failures. This raises a fundamental question for embodied agents: when should the agent reason, and when should it act? In this work, we propose RARRL (Resource-Aware Reasoning via Reinforcement Learning), a hierarchical framework for resource-aware orchestration of embodied agents. Rather than learning low-level control policies, RARRL learns a high-level orchestration policy that operates at the agent's decision-making layer. This policy enables the agent to adaptively determine whether to invoke reasoning, which reasoning role to employ, and how much computational budget to allocate based on current observations, execution history, and remaining resources. Extensive experiments, including evaluations with empirical latency profiles derived from the ALFRED benchmark, show that RARRL consistently improves task success rates while reducing execution latency and enhancing robustness compared with fixed or heuristic reasoning strategies. These results demonstrate that adaptive reasoning control is essential for building reliable and efficient embodied robotic agents.