Learning Multi-Modal Whole-Body Control for Real-World Humanoid Robots
arXiv cs.RO / 4/23/2026
📰 NewsDeveloper Stack & InfrastructureModels & Research
Key Points
- The paper introduces the Masked Humanoid Controller (MHC), a learned whole-body control method for humanoid robots that uses masked target trajectories over selected robot state variables as a unified command interface.
- The approach lets high-level systems specify diverse behaviors—such as footstep plans, partial-body mimicry, motion-capture-driven motions, video retargeting, and joystick teleoperation—in a flexible format.
- MHC is trained in simulation with a curriculum spanning multiple input modalities, aiming to robustly execute partially specified behaviors while preserving balance and disturbance rejection.
- The authors evaluate MHC in both simulation and on a real-world Digit V3 humanoid, finding that one controller can handle multiple command types via the same representation.
- Overall, the work targets the longstanding challenge of providing a single interface that can command many whole-body behaviors without redesigning controllers for each modality.
Related Articles

Trajectory Forecasts in Unknown Environments Conditioned on Grid-Based Plans
Dev.to

Elevating Austria: Google invests in its first data center in the Alps.
Google Blog

OpenAI Just Named It Workspace Agents. We Open-Sourced Our Lark Version Six Months Ago
Dev.to

GPT Image 2 Subject-Lock Editing: A Practical Guide to input_fidelity
Dev.to

AI Tutor That Works Offline — Study Anywhere with EaseLearn AI
Dev.to