Generative Simulation for Policy Learning in Physical Human-Robot Interaction
arXiv cs.RO / 4/13/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper proposes a zero-shot “text2sim2real” generative simulation framework for physical human-robot interaction that synthesizes diverse scenarios from natural-language prompts.
- It uses LLMs and VLMs to procedurally generate soft-body human models, scene layouts, and robot motion trajectories for assistive tasks.
- The framework enables large-scale synthetic demonstration collection and trains vision-based imitation learning policies using segmented point clouds.
- Experiments via a user study on scratching and bathing show the learned policies achieve zero-shot sim-to-real transfer with success rates above 80% and robustness to unscripted human motion.
- The authors position this as the first generative simulation pipeline that automates simulation environment synthesis, synthetic data generation, and policy learning for pHRI.
Related Articles

Black Hat Asia
AI Business

I built the missing piece of the MCP ecosystem
Dev.to

When Agents Go Wrong: AI Accountability and the Payment Audit Trail
Dev.to

Google Gemma 4 Review 2026: The Open Model That Runs Locally and Beats Closed APIs
Dev.to

OpenClaw Deep Dive Guide: Self-Host Your Own AI Agent on Any VPS (2026)
Dev.to