Interval POMDP Shielding for Imperfect-Perception Agents

arXiv cs.AI / 4/23/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses how autonomous agents can make unsafe decisions when learned perception misclassifies sensor readings, and introduces a “shielding” approach to prevent unsafe actions.
  • It models known system dynamics with perception uncertainty learned from finite labeled data by building confidence intervals for perception outcome probabilities.
  • The authors formulate the problem as a finite Interval Partially Observable Markov Decision Process (Interval POMDP) with discrete states and actions, and develop an algorithm to compute conservative belief sets consistent with past observations.
  • A runtime shield is proposed that provides a finite-horizon safety guarantee: assuming the true perception uncertainty rates fall within the learned intervals, every action allowed by the shield meets a stated lower bound on safety with high probability over the training data.
  • Experiments across four case studies show improved safety compared with state-of-the-art baselines, including variants derived from their method.

Abstract

Autonomous systems that rely on learned perception can make unsafe decisions when sensor readings are misclassified. We study shielding for this setting: given a proposed action, a shield blocks actions that could violate safety. We consider the common case where system dynamics are known but perception uncertainty must be estimated from finite labeled data. From these data we build confidence intervals for the probabilities of perception outcomes and use them to model the system as a finite Interval Partially Observable Markov Decision Process with discrete states and actions. We then propose an algorithm to compute a conservative set of beliefs over the underlying state that is consistent with the observations seen so far. This enables us to construct a runtime shield that comes with a finite-horizon guarantee: with high probability over the training data, if the true perception uncertainty rates lie within the learned intervals, then every action admitted by the shield satisfies a stated lower bound on safety. Experiments on four case studies show that our shielding approach (and variants derived from it) improves the safety of the system over state-of-the-art baselines.