VOFA: Visual Object Goal Pushing with Force-Adaptive Control for Humanoids
arXiv cs.RO / 5/5/2026
📰 NewsDeveloper Stack & InfrastructureModels & Research
Key Points
- The paper introduces VOFA, a humanoid loco-manipulation system that can push heavy objects to arbitrary goal locations using onboard egocentric visual perception rather than privileged state information.
- VOFA uses a two-level hierarchical design: a high-level visuomotor policy that converts noisy visual observations into goal-conditioned commands and a low-level force-adaptive whole-body controller to handle unknown object mass and ground friction.
- The approach is designed to remain robust under real-world uncertainties, including noisy sensing and actuation errors, by operating in a real-time closed feedback loop.
- Experiments on the Booster T1 humanoid robot show strong results, with over 90% success in simulation and over 80% success in real-world trials.
- VOFA can push objects up to 17 kg—more than half of the Booster T1’s body weight—demonstrating its capability for physically demanding material-handling tasks.
Related Articles

Why Retail Chargeback Recovery Could Be AgentHansa's First Real PMF
Dev.to
Struggling with Qwen3.6 27B / 35B locally (3090) slow responses, breaking code looking for better setup + auto model switching
Reddit r/LocalLLaMA

Last Week in AI #340 - OpenAI vs Musk + Microsoft, DeepSeek v4, Vision Banana
Last Week in AI

Trying to train tiny LLMs on length constrained reddit posts summarization task using GRPO on 3xMac Minis - updates!
Reddit r/LocalLLaMA

Uber Shares What Happens When 1.500 AI Agents Hit Production
Reddit r/artificial