Chasing Autonomy: Dynamic Retargeting and Control Guided RL for Performant and Controllable Humanoid Running
arXiv cs.RO / 3/30/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes a pipeline that dynamically retargets a humanoid’s motion using an optimization routine with hard constraints, producing periodic reference libraries from a single human demonstration.
- It finds that both the choice of reference motion and the reward design significantly affect velocity tracking, with a goal-conditioned, control-guided reward that follows dynamically optimized human data delivering the best performance.
- The authors deploy the resulting RL policy on a Unitree G1 robot, demonstrating fast and endurance-oriented running up to 3.3 m/s and hundreds of meters of real-world traversal.
- The work also showcases locomotion controllability by integrating the controller into a perception-and-planning autonomy stack for outdoor obstacle avoidance while running.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.




