CORA: Conformal Risk-Controlled Agents for Safeguarded Mobile GUI Automation
arXiv cs.LG / 4/13/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- CORA is proposed as a post-policy, pre-action safeguarding framework for VLM-powered autonomous mobile GUI agents, focusing on statistically guaranteed reduction of harmful executed actions.
- The method trains a Guardian model to estimate action-conditional risk and uses Conformal Risk Control to create a calibrated execute/abstain decision boundary aligned with a user-specified risk budget.
- Rejected (high-risk) actions are routed to a trainable Diagnostician that performs multimodal reasoning to recommend interventions such as confirm, reflect, or abort, aiming to reduce user burden.
- A Goal-Lock mechanism is introduced to anchor risk assessment to clarified, frozen user intent, helping resist visual injection attacks.
- The paper also introduces the Phone-Harm benchmark with step-level harm labels under real-world mobile settings and reports experimental results showing improved safety–helpfulness–interruption trade-offs, with code and benchmarks published online.
Related Articles

Black Hat Asia
AI Business

I built the missing piece of the MCP ecosystem
Dev.to

When Agents Go Wrong: AI Accountability and the Payment Audit Trail
Dev.to

Google Gemma 4 Review 2026: The Open Model That Runs Locally and Beats Closed APIs
Dev.to

OpenClaw Deep Dive Guide: Self-Host Your Own AI Agent on Any VPS (2026)
Dev.to