Hypernetwork-Conditioned Reinforcement Learning for Robust Control of Fixed-Wing Aircraft under Actuator Failures

arXiv cs.LG / 4/8/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces a reinforcement learning path-following controller for a fixed-wing small UAV that remains robust under actuator failures.
  • It uses a hypernetwork to condition the policy on a parameterized representation of actuator fault modes, enabling adaptation to different failure conditions.
  • The method explores parameter-efficient conditioning via FiLM and LoRA and trains the resulting policy with proximal policy optimization (PPO).
  • Experiments on high-fidelity 6-DOF fixed-wing simulations show that hypernetwork-conditioned policies outperform standard MLP-based policies in robustness.
  • The approach is reported to generalize to time-varying actuator failure modes that were not present during training.

Abstract

This paper presents a reinforcement learning-based path-following controller for a fixed-wing small uncrewed aircraft system (sUAS) that is robust to certain actuator failures. The controller is conditioned on a parameterization of actuator faults using hypernetwork-based adaptation. We consider parameter-efficient formulations based on Feature-wise Linear Modulation (FiLM) and Low-Rank Adaptation (LoRA), trained using proximal policy optimization. We demonstrate that hypernetwork-conditioned policies can improve robustness compared to standard multilayer perceptron policies. In particular, hypernetwork-conditioned policies generalize effectively to time-varying actuator failure modes not encountered during training. The approach is validated through high-fidelity simulations, using a realistic six-degree-of-freedom fixed-wing aircraft model.