RiskWebWorld: A Realistic Interactive Benchmark for GUI Agents in E-commerce Risk Management

arXiv cs.AI / 4/16/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • RiskWebWorld is presented as a realistic interactive benchmark designed specifically to evaluate GUI agents for high-stakes e-commerce risk management, extending beyond previously benign consumer web settings.
  • The benchmark includes 1,513 tasks drawn from production risk-control pipelines across eight domains, modeling real operational difficulties such as uncooperative websites and partial environmental hijackments.
  • To enable scalable testing and training, the authors provide a Gymnasium-compliant evaluation infrastructure that separates policy planning from environment mechanics.
  • Experiments show a large performance gap: top-tier generalist models reach 49.1% task success, while specialized open-weight GUI models perform near-total failure, suggesting model scale is more important than zero-shot interface grounding for long-horizon work.
  • Agentic reinforcement learning using the provided infrastructure improves open-source models by 16.2%, demonstrating the benchmark’s usefulness as a testbed for building more reliable “digital workers.”

Abstract

Graphical User Interface (GUI) agents show strong capabilities for automating web tasks, but existing interactive benchmarks primarily target benign, predictable consumer environments. Their effectiveness in high-stakes, investigative domains such as authentic e-commerce risk management remains underexplored. To bridge this gap, we present RiskWebWorld, the first highly realistic interactive benchmark for evaluating GUI agents in e-commerce risk management. RiskWebWorld features 1,513 tasks sourced from production risk-control pipelines across 8 core domains, and captures the authentic challenges of risk operations on uncooperative websites, partially environmental hijackments. To support scalable evaluation and agentic reinforcement learning (RL), we further build a Gymnasium-compliant infrastructure that decouples policy planning from environment mechanics. Our evaluation across diverse models reveals a dramatic capability gap: top-tier generalist models achieve 49.1% success, while specialized open-weights GUI models lag at near-total failure. This highlights that foundation model scale currently matters more than zero-shot interface grounding in long-horizon professional tasks. We also demonstrate the viability of our infrastructure through agentic RL, which improves open-source models by 16.2%. These results position RiskWebWorld as a practical testbed for developing robust digital workers.