Rule-based High-Level Coaching for Goal-Conditioned Reinforcement Learning in Search-and-Rescue UAV Missions Under Limited-Simulation Training

arXiv cs.RO / 4/30/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper introduces a hierarchical UAV decision-making framework for search-and-rescue scenarios that pairs a fixed, rule-based high-level advisor with an online goal-conditioned reinforcement learning (RL) low-level controller.
  • The high-level advisor is precompiled into deterministic, interpretable rules from a structured task specification, enabling mission- and safety-aware guidance including recommended/avoided actions and arbitration weights.
  • The low-level RL component is trained online under limited-simulation conditions (including a strict no-pretraining regime) using task-defined dense rewards and enhanced experience replay that leverages rule-derived metadata.
  • Experiments on battery-aware multi-goal delivery and moving-target delivery in obstacle-rich environments show improved early safety and sample efficiency, mainly by reducing collision-related terminations, while still allowing online adaptation to scenario dynamics.

Abstract

This paper presents a hierarchical decision-making framework for unmanned aerial vehicle (UAV) missions motivated by search-and-rescue (SAR) scenarios under limited simulation training. The framework combines a fixed rule-based high-level advisor with an online goal-conditioned low-level reinforcement learning (RL) controller. To stress-test early adaptation, we also consider a strict no-pretraining deployment regime. The high-level advisor is defined offline from a structured task specification and compiled into deterministic rules. It provides interpretable mission- and safety-aware guidance through recommended actions, avoided actions, and regime-dependent arbitration weights. The low-level controller learns online from task-defined dense rewards and reuses experience through a mode-aware prioritized replay mechanism augmented with rule-derived metadata. We evaluate the framework on two tasks: battery-aware multi-goal delivery and moving-target delivery in obstacle-rich environments. Across both tasks, the proposed method improves early safety and sample efficiency primarily by reducing collision terminations, while preserving the ability to adapt online to scenario-specific dynamics.