Robot Learning from Human Videos: A Survey
arXiv cs.CV / 5/1/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The survey identifies scaling robot data as a major bottleneck in embodied AI and robotics, and highlights human-video-based learning as a promising approach to alleviate it.
- It reviews foundational policy learning concepts in robotics and the key interfaces for incorporating human videos into robot learning pipelines.
- The paper proposes a hierarchical taxonomy for transferring human videos into robot skills, organized by task-, observation-, and action-oriented pathways, and analyzes how these methods relate across data setups and learning paradigms.
- It examines data foundations, including widely used human-video datasets, video generation methods, and large-scale statistics on dataset creation and utilization trends.
- The survey concludes by outlining core challenges and limitations of the field and suggesting directions for future research.
- The work also provides a curated, up-to-date reading list via a linked GitHub repository.
Related Articles

Why Autonomous Coding Agents Keep Failing — And What Actually Works
Dev.to

Why Enterprise AI Pilots Fail
Dev.to

The PDF Feature Nobody Asked For (That I Use Every Day)
Dev.to

How to Fix OpenClaw Tool Calling Issues
Dev.to

Mistral's new flagship Medium 3.5 folds chat, reasoning, and code into one model
THE DECODER