Learning Whole-Body Humanoid Locomotion via Motion Generation and Motion Tracking

arXiv cs.RO / 4/21/2026

📰 NewsDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper tackles whole-body humanoid locomotion, which is difficult due to high-dimensional control, morphological instability, and the need to adapt in real time to changing terrain using onboard perception.
  • It introduces a framework that pairs a diffusion-based motion generation model (trained on retargeted human motions) with an RL-trained whole-body motion tracker to produce terrain-aware reference motions.
  • To handle imperfect or noisy generated references, the authors fine-tune the tracker in a closed-loop setup while keeping the motion generator frozen, improving robustness.
  • The approach enables directional goal-reaching with terrain-aware whole-body adaptation and is validated on a Unitree G1 humanoid robot using onboard perception and computation across boxes, hurdles, stairs, and mixed terrains.
  • Quantitative experiments show that online motion generation and closed-loop fine-tuning of the motion tracker improve generalization and robustness compared with alternatives that rely more heavily on replaying fixed reference motions.

Abstract

Whole-body humanoid locomotion is challenging due to high-dimensional control, morphological instability, and the need for real-time adaptation to various terrains using onboard perception. Directly applying reinforcement learning (RL) with reward shaping to humanoid locomotion often leads to lower-body-dominated behaviors, whereas imitation-based RL can learn more coordinated whole-body skills but is typically limited to replaying reference motions without a mechanism to adapt them online from perception for terrain-aware locomotion. To address this gap, we propose a whole-body humanoid locomotion framework that combines skills learned from reference motions with terrain-aware adaptation. We first train a diffusion model on retargeted human motions for real-time prediction of terrain-aware reference motions. Concurrently, we train a whole-body reference tracker with RL using this motion data. To improve robustness under imperfectly generated references, we further fine-tune the tracker with a frozen motion generator in a closed-loop setting. The resulting system supports directional goal-reaching control with terrain-aware whole-body adaptation, and can be deployed on a Unitree G1 humanoid robot with onboard perception and computation. The hardware experiments demonstrate successful traversal over boxes, hurdles, stairs, and mixed terrain combinations. Quantitative results further show the benefits of incorporating online motion generation and fine-tuning the motion tracker for improved generalization and robustness.