Robustness Analysis of POMDP Policies to Observation Perturbations

arXiv cs.AI / 4/25/2026

📰 NewsModels & Research

Key Points

  • The paper examines how POMDP policies degrade when the deployed observation model deviates from the nominal model due to real-world effects like calibration drift and sensor degradation.
  • It formalizes a “Policy Observation Robustness Problem” that computes the maximum tolerable observation-model deviation while ensuring the policy value stays above a chosen threshold.
  • Two robustness variants are analyzed: a sticky deviation model (dependent on state/actions) and a non-sticky model (history-dependent), with the inner optimization shown to be monotonic in deviation size.
  • The authors develop efficient solution methods by casting the problem as bi-level optimization and using root-finding in the outer loop; for non-sticky cases with finite-state controllers (FSCs), they show it suffices to consider observations tied to FSC nodes rather than full histories.
  • They introduce “Robust Interval Search,” proving soundness and convergence, and provide complexity results (polynomial for non-sticky, up to exponential for sticky) plus experiments scaling to POMDPs with tens of thousands of states and robotics/operations-research case studies.

Abstract

Policies for Partially Observable Markov Decision Processes (POMDPs) are often designed using a nominal system model. In practice, this model can deviate from the true system during deployment due to factors such as calibration drift or sensor degradation, leading to unexpected performance degradation. This work studies policy robustness against deviations in the POMDP observation model. We introduce the Policy Observation Robustness Problem: to determine the maximum tolerable deviation in a POMDP's observation model that guarantees the policy's value remains above a specified threshold. We analyze two variants: the sticky variant, where deviations are dependent on state and actions, and the non-sticky variant, where they can be history-dependent. We show that the Policy Observation Robustness Problem can be formulated as a bi-level optimization problem in which the inner optimization is monotonic in the size of the observation deviation. This enables efficient solutions using root-finding algorithms in the outer optimization. For the non-sticky variant, we show that when policies are represented with finite-state controllers (FSCs) it is sufficient to consider observations which depend on nodes in the FSC rather than full histories. We present Robust Interval Search, an algorithm with soundness and convergence guarantees, for both the sticky and non-sticky variants. We show this algorithm has polynomial time complexity in the non-sticky variant and at most exponential time complexity in the sticky variant. We provide experimental results validating and demonstrating the scalability of implementations of Robust Interval Search to POMDP problems with tens of thousands of states. We also provide case studies from robotics and operations research which demonstrate the practical utility of the problem and algorithms.