Decentralized End-to-End Multi-AAV Pursuit Using Predictive Spatio-Temporal Observation via Deep Reinforcement Learning

arXiv cs.RO / 3/26/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper presents a decentralized end-to-end MARL framework that converts raw LiDAR inputs directly into continuous control commands for autonomous aerial swarms in cluttered, partially observed environments.
  • It introduces Predictive Spatio-Temporal Observation (PSTO), an egocentric fixed-resolution grid representation that unifies obstacle geometry, predicted adversarial intent, and teammate motion to better handle perceptual uncertainty.
  • A single decentralized policy learned with PSTO is designed to support multiple cooperative behaviors, including navigating static obstacles, intercepting dynamic targets, and maintaining encirclement.
  • Simulation results show improved capture efficiency and success rates versus prior learning-based methods that depend on privileged obstacle/state information.
  • The authors report that the same unified policy transfers across different team sizes without retraining and is validated through fully autonomous outdoor quadrotor swarm experiments using only onboard sensing and computation.

Abstract

Decentralized cooperative pursuit in cluttered environments is challenging for autonomous aerial swarms, especially under partial and noisy perception. Existing methods often rely on abstracted geometric features or privileged ground-truth states, and therefore sidestep perceptual uncertainty in real-world settings. We propose a decentralized end-to-end multi-agent reinforcement learning (MARL) framework that maps raw LiDAR observations directly to continuous control commands. Central to the framework is the Predictive Spatio-Temporal Observation (PSTO), an egocentric grid representation that aligns obstacle geometry with predictive adversarial intent and teammate motion in a unified, fixed-resolution projection. Built on PSTO, a single decentralized policy enables agents to navigate static obstacles, intercept dynamic targets, and maintain cooperative encirclement. Simulations demonstrate that the proposed method achieves superior capture efficiency and competitive success rates compared to state-of-the-art learning-based approaches relying on privileged obstacle information. Furthermore, the unified policy scales seamlessly across different team sizes without retraining. Finally, fully autonomous outdoor experiments validate the framework on a quadrotor swarm relying on only onboard sensing and computing.