Dyna-Style Safety Augmented Reinforcement Learning: Staying Safe in the Face of Uncertainty

arXiv cs.LG / 4/29/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses a core challenge in reinforcement learning: ensuring safety during training, particularly when system dynamics are unknown and environments are high-dimensional.
  • It introduces Dyna-style Safety Augmented Reinforcement Learning (Dyna-SAuR), which jointly learns a scalable safety filter and a control policy using an uncertainty-aware learned dynamics model.
  • The learned safety filter is designed to actively steer the agent away from failure modes and regions with high uncertainty, improving safety without overly conservative restrictions.
  • By leveraging improved learned models, Dyna-SAuR can expand the set of “safe and certain” states, thereby reducing the conservatism typical of safety filters.
  • Experiments on CartPole and MuJoCo Walker show that Dyna-SAuR reduces failures by about two orders of magnitude versus state-of-the-art approaches.

Abstract

Safety remains an open problem in reinforcement learning (RL), especially during training. While safety filters are promising to address safe exploration, they are generally poorly suited for high-dimensional systems with unknown dynamics. We propose Dyna-style Safety Augmented Reinforcement Learning (Dyna-SAuR), a novel algorithm that learns both a scalable safety filter and a control policy using a learned uncertainty-aware dynamics model, while requiring minimal domain knowledge. The filter avoids failures and high uncertainty regions. Thus, better models expand the set of safe and certain states, reducing filter conservatism. We present the effectiveness of Dyna-SAuR on goal-reaching CartPole as well as MuJoCo Walker, reducing failures compared to state-of-the-art methods by 2 orders of magnitude.