CaP-X: A Framework for Benchmarking and Improving Coding Agents for Robot Manipulation

arXiv cs.RO / 3/25/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces CaP-X, an open-access framework to benchmark and improve “Code-as-Policy” coding agents for embodied robot manipulation.
  • Its core component, CaP-Gym, provides an interactive environment where agents control robots by synthesizing and executing programs that combine perception and control primitives.
  • Using CaP-Bench, the authors evaluate 12 frontier language/vision-language models and find performance rises with human-crafted abstractions but drops when those priors are removed, highlighting dependence on designer scaffolding.
  • The study shows robustness can be improved via scaling test-time computation (e.g., multi-turn interaction, structured execution feedback, visual differencing, skill synthesis, and ensembling) and proposes CaP-Agent0 as a training-free method achieving human-level reliability in multiple tasks.
  • It also proposes CaP-RL, demonstrating that reinforcement learning with verifiable rewards improves success rates and enables better sim-to-real transfer with a small gap.

Abstract

"Code-as-Policy" considers how executable code can complement data-intensive Vision-Language-Action (VLA) methods, yet their effectiveness as autonomous controllers for embodied manipulation remains underexplored. We present CaP-X, an open-access framework for systematically studying Code-as-Policy agents in robot manipulation. At its core is CaP-Gym, an interactive environment in which agents control robots by synthesizing and executing programs that compose perception and control primitives. Building on this foundation, CaP-Bench evaluates frontier language and vision-language models across varying levels of abstraction, interaction, and perceptual grounding. Across 12 models, CaP-Bench reveals a consistent trend: performance improves with human-crafted abstractions but degrades as these priors are removed, exposing a dependence on designer scaffolding. At the same time, we observe that this gap can be mitigated through scaling agentic test-time computation--through multi-turn interaction, structured execution feedback, visual differencing, automatic skill synthesis, and ensembled reasoning--substantially improves robustness even when agents operate over low-level primitives. These findings allow us to derive CaP-Agent0, a training-free framework that recovers human-level reliability on several manipulation tasks in simulation and on real embodiments. We further introduce CaP-RL, showing reinforcement learning with verifiable rewards improves success rates and transfers from sim2real with minimal gap. Together, CaP-X provides a principled, open-access platform for advancing embodied coding agents.