Agent-Driven Autonomous Reinforcement Learning Research: Iterative Policy Improvement for Quadruped Locomotion

arXiv cs.RO / 3/31/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper presents a case study of agent-driven autonomous reinforcement learning for quadruped locomotion, where an agent executes most of the experiment loop (coding, debugging, reward/terrain edits, job running, monitoring, and proposing follow-up experiments).
  • Across 70+ experiments in 14 iterative waves on a DHAV1 12-DoF quadruped in Isaac Lab, the system improved from early rough-terrain mean reward (~7) to a best Wave 12 result with velocity error 0.263 and 97% timeout over 2000 iterations, reproducible on multiple GPUs.
  • The study documents concrete research decisions that the agent made, including diagnosing simulator issues (e.g., PhysX deadlocks), porting and adjusting reward terms from reference implementations, and engineering fixes for Isaac Sim import/bootstrapping problems.
  • It also highlights practical guardrails and pivots (reducing environment counts for faster diagnosis, terminating hung runs, and redirecting effort when terrain outcomes repeatedly collapsed to 0.0).
  • Compared with AutoResearch, the work emphasizes a more failure-prone robotics RL environment with multi-GPU experiment management and simulator-specific constraints, positioning the contribution as empirical/archival rather than a fully self-starting system.

Abstract

This paper documents a case study in agent-driven autonomous reinforcement learning research for quadruped locomotion. The setting was not a fully self-starting research system. A human provided high-level directives through an agentic coding environment, while an agent carried out most of the execution loop: reading code, diagnosing failures, editing reward and terrain configurations, launching and monitoring jobs, analyzing intermediate metrics, and proposing the next wave of experiments. Across more than 70 experiments organized into fourteen waves on a DHAV1 12-DoF quadruped in Isaac Lab, the agent progressed from early rough-terrain runs with mean reward around 7 to a best logged Wave 12 run, exp063, with velocity error 0.263 and 97\% timeout over 2000 iterations, independently reproduced five times across different GPUs. The archive also records several concrete autonomous research decisions: isolating PhysX deadlocks to terrain sets containing boxes and stair-like primitives, porting four reward terms from openly available reference implementations \cite{deeprobotics, rlsar}, correcting Isaac Sim import and bootstrapping issues, reducing environment count for diagnosis, terminating hung runs, and pivoting effort away from HIM after repeated terrain=0.0 outcomes. Relative to the AutoResearch paradigm \cite{autoresearch}, this case study operates in a more failure-prone robotics RL setting with multi-GPU experiment management and simulator-specific engineering constraints. The contribution is empirical and documentary: it shows that an agent can materially execute the iterative RL research loop in this domain with limited human intervention, while also making clear where human direction still shaped the agenda.