Augmented Lagrangian Multiplier Network for State-wise Safety in Reinforcement Learning

arXiv cs.LG / 5/4/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes ALaM (Augmented Lagrangian Multiplier Network) to learn state-wise safety constraints in reinforcement learning by using a neural multiplier network that varies by state.
  • It argues that naive dual gradient ascent over state-dependent multipliers leads to severe training oscillations due to the instability of dual ascent combined with neural generalization across states.
  • ALaM stabilizes learning by adding a quadratic penalty in the augmented Lagrangian to improve local convexity and by training the multiplier network via supervised regression toward dual targets.
  • The authors provide theoretical guarantees that the multipliers converge and that the method recovers the optimal constrained policy, and they instantiate the approach as SAC-ALaM by integrating with soft actor-critic.
  • Experiments show SAC-ALaM improves over prior safe RL baselines on both safety and return, while also producing well-calibrated multipliers for risk identification.

Abstract

Safety is a primary challenge in real-world reinforcement learning (RL). Formulating safety requirements as state-wise constraints has become a prominent paradigm. Handling state-wise constraints with the Lagrangian method requires a distinct multiplier for every state, necessitating neural networks to approximate them as a multiplier network. However, applying standard dual gradient ascent to multiplier networks induces severe training oscillations. This is because the inherent instability of dual ascent is exacerbated by network generalization -- local overshoots and delayed updates propagate to adjacent states, further amplifying policy fluctuations. Existing stabilization techniques are designed for scalar multipliers, which are inadequate for state-dependent multiplier networks. To address this challenge, we propose an augmented Lagrangian multiplier network (ALaM) framework for stable learning of state-wise multipliers. ALaM consists of two key components. First, a quadratic penalty is introduced into the augmented Lagrangian to compensate for delayed multiplier updates and establish the local convexity near the optimum, thereby mitigating policy oscillations. Second, the multiplier network is trained via supervised regression toward a dual target, which stabilizes training and promotes convergence. Theoretically, we show that ALaM guarantees multiplier convergence and thus recovers the optimal policy of the constrained problem. Building on this framework, we integrate soft actor-critic (SAC) with ALaM to develop the SAC-ALaM algorithm. Experiments demonstrate that SAC-ALaM outperforms state-of-the-art safe RL baselines in both safety and return, while also stabilizing training dynamics and learning well-calibrated multipliers for risk identification.