Learning Safe-Stoppability Monitors for Humanoid Robots

arXiv cs.RO / 3/25/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • Humanoid emergency stops can’t simply cut power, because abrupt shutdown may destabilize the robot; instead, the robot must switch to a predefined fallback controller that reaches a minimum-risk condition.
  • The paper formalizes this as a policy-dependent “safe-stoppability” problem, defining which states are safe for executing an emergency stop for a given robot policy.
  • It introduces PRISM, a simulation-driven framework that learns a neural state-level stoppability monitor and refines its decision boundary using importance sampling to efficiently cover rare, safety-critical scenarios.
  • The authors report improved data efficiency and fewer false-safe predictions under a fixed simulation budget, and they validate sim-to-real transfer by running the pretrained monitor on a real humanoid robot.
  • By modeling safety as stoppability, the approach aims to enable proactive safety monitoring and more scalable certification of fail-safe behaviors for humanoids.

Abstract

Emergency stop (E-stop) mechanisms are the de facto standard for robot safety. However, for humanoid robots, abruptly cutting power can itself cause catastrophic failures; instead, an emergency stop must execute a predefined fallback controller that preserves balance and drives the robot toward a minimum-risk condition. This raises a critical question: from which states can a humanoid robot safely execute such a stop? In this work, we formalize emergency stopping for humanoids as a policy-dependent safe-stoppability problem and use data-driven approaches to characterize the safe-stoppable envelope. We introduce PRISM (Proactive Refinement of Importance-sampled Stoppability Monitor), a simulation-driven framework that learns a neural predictor for state-level stoppability. PRISM iteratively refines the decision boundary using importance sampling, enabling targeted exploration of rare but safety-critical states. This targeted exploration significantly improves data efficiency while reducing false-safe predictions under a fixed simulation budget. We further demonstrate sim-to-real transfer by deploying the pretrained monitor on a real humanoid platform. Results show that modeling safety as policy-dependent stoppability enables proactive safety monitoring and supports scalable certification of fail-safe behaviors for humanoid robots.