Faithful Mobile GUI Agents with Guided Advantage Estimator

arXiv cs.AI / 5/5/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that vision-language GUI agents can act unfaithfully by using memorized shortcuts instead of grounding actions in visible screen evidence or user instructions.
  • It introduces Faithful-Agent, a faithfulness-first framework that reshapes GUI interaction to emphasize evidence-groundedness and internal consistency.
  • Faithful-Agent uses a two-stage training pipeline: a faithfulness-oriented SFT stage to encourage abstention when evidence is perturbed, followed by an RFT stage to further boost faithfulness.
  • The RFT stage adds a guided advantage estimator (GuAE) based on GRPO, designed to prevent advantage collapse in low-variance rollout groups when GUI rewards are sparse.
  • With an additional thought-action consistency reward, the Stage-II method raises the Trap SR from 13.88% to 80.21% versus the baseline while maintaining strong performance on general instruction following.

Abstract

Vision-language model based graphical user interface (GUI) agents have shown strong interaction capabilities. However, they often behave unfaithfully, relying on memorized shortcuts rather than grounding actions in displayed screen evidence or user instructions. To address this, we propose Faithful-Agent, a faithfulness-first framework that reformulates GUI interaction to prioritize evidence groundedness and internal consistency. Faithful-Agent employs a two-stage pipeline: (i) a faithfulness-oriented SFT stage to instill abstainment behaviors under evidence perturbations; (ii) an RFT stage that further amplifies faithfulness by introducing the guided advantage estimator (GuAE), an anchor-based and variance-adaptive advantage tempering mechanism built upon GRPO. GuAE prevents advantage collapse in low-variance rollout groups under sparse GUI rewards, and with a thought-action consistency reward, Faithful-Agent (Stage II) elevates the Trap SR from 13.88\% to 80.21\% relative to the baseline, while preserving robust general instruction-following performance.