AI Navigate

OS-Themis: A Scalable Critic Framework for Generalist GUI Rewards

arXiv cs.AI / 3/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • OS-Themis is a scalable multi-agent critic framework for reinforcement learning that decomposes GUI task trajectories into verifiable milestones to improve reward quality.
  • It employs a review mechanism to audit the evidence chain before reaching a final verdict, reducing reliance on a single judge.
  • The work introduces OmniGUIRewardBench (OGRBench), a cross-platform benchmark for GUI outcome rewards to facilitate evaluation under OS-Themis.
  • Experimental results on AndroidWorld show OS-Themis yields about a 10.3% improvement in online RL training and a 6.9% gain in trajectory validation within a self-training loop, highlighting its potential to advance GUI agent evolution.

Abstract

Reinforcement Learning (RL) has the potential to improve the robustness of GUI agents in stochastic environments, yet training is highly sensitive to the quality of the reward function. Existing reward approaches struggle to achieve both scalability and performance. To address this, we propose OS-Themis, a scalable and accurate multi-agent critic framework. Unlike a single judge, OS-Themis decomposes trajectories into verifiable milestones to isolate critical evidence for decision making and employs a review mechanism to strictly audit the evidence chain before making the final verdict. To facilitate evaluation, we further introduce OmniGUIRewardBench (OGRBench), a holistic cross-platform benchmark for GUI outcome rewards, where all evaluated models achieve their best performance under OS-Themis. Extensive experiments on AndroidWorld show that OS-Themis yields a 10.3% improvement when used to support online RL training, and a 6.9% gain when used for trajectory validation and filtering in the self-training loop, highlighting its potential to drive agent evolution.