Safe Control using Learned Safety Filters and Adaptive Conformal Inference

arXiv cs.RO / 4/21/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces Adaptive Conformal Filtering (ACoFi), which combines learned Hamilton–Jacobi reachability-based safety filters with adaptive conformal inference for control systems that use an unsafe nominal policy.
  • ACoFi dynamically updates its switching criteria by using the observed prediction errors and by estimating an uncertainty range for the nominal policy’s safety values.
  • When the estimated range suggests the nominal action may be unsafe, the safety filter switches to a learned safe policy rather than relying on a fixed threshold.
  • The method provides a “soft” safety guarantee: the rate of incorrect uncertainty quantification for the nominal policy’s predicted safety is asymptotically upper bounded by a user-specified parameter.
  • Experiments in a Dubins car simulation and Safety Gymnasium show ACoFi improves over a fixed-threshold baseline, producing higher learned safety values and fewer safety violations, particularly under out-of-distribution conditions.

Abstract

Safety filters have been shown to be effective tools to ensure the safety of control systems with unsafe nominal policies. To address scalability challenges in traditional synthesis methods, learning-based approaches have been proposed for designing safety filters for systems with high-dimensional state and control spaces. However, the inevitable errors in the decisions of these models raise concerns about their reliability and the safety guarantees they offer. This paper presents Adaptive Conformal Filtering (ACoFi), a method that combines learned Hamilton-Jacobi reachability-based safety filters with adaptive conformal inference. Under ACoFi, the filter dynamically adjusts its switching criteria based on the observed errors in its predictions of the safety of actions. The range of possible safety values of the nominal policy's output is used to quantify uncertainty in safety assessment. The filter switches from the nominal policy to the learned safe one when that range suggests it might be unsafe. We show that ACoFi guarantees that the rate of incorrectly quantifying uncertainty in the predicted safety of the nominal policy is asymptotically upper bounded by a user-defined parameter. This gives a soft safety guarantee rather than a hard safety guarantee. We evaluate ACoFi in a Dubins car simulation and a Safety Gymnasium environment, empirically demonstrating that it significantly outperforms the baseline method that uses a fixed switching threshold by achieving higher learned safety values and fewer safety violations, especially in out-of-distribution scenarios.