EgoLive: A Large-Scale Egocentric Dataset from Real-World Human Tasks

arXiv cs.RO / 4/28/2026

📰 NewsSignals & Early TrendsModels & Research

Key Points

  • The paper introduces EgoLive, a large-scale, high-quality egocentric dataset aimed at improving robot manipulation learning amid limited dataset availability.
  • EgoLive claims three technical advantages: the largest open-source annotated egocentric dataset for real-world task routines so far, state-of-the-art data quality from a customized head-mounted capture setup, and comprehensive high-precision multimodal annotations.
  • Unlike many existing approaches (e.g., teleoperation or universal manipulation interfaces), EgoLive is collected exclusively in unconstrained real-world settings to enhance scalability and ecological validity.
  • The dataset includes vertical-field human working data spanning home services, retail, and other practical work scenarios, targeting greater diversity for more generalizable robotic models.
  • The authors position EgoLive as a resource to accelerate breakthroughs and support the real-world deployment of robot systems by providing scalable training data.

Abstract

The advancement of robot learning is currently hindered by the scarcity of large-scale, high-quality datasets. While established data collection methods such as teleoperation and universal manipulation interfaces dominate current datasets, they suffer from inherent limitations in scalability and real-world deployability. Human egocentric video collection, by contrast, has emerged as a promising approach to enable scalable, natural and in-the-wild data collection. As such, we present EgoLive, a large-scale, high-quality egocentric dataset designed explicitly for robot manipulation learning. EgoLive establishes three distinctive technical advantages over existing egocentric datasets: first, it represents the largest open-source annotated egocentric dataset focused on real-world task-oriented human routines to date; second, it delivers leading data quality via a customized head-mounted capture device and comprehensive high-precision multi-modal annotations; third, all data is collected exclusively in unconstrained real-world scenarios and encompasses vertical field human working data, including home service, retail, and other practical work scenarios, providing superior diversity and ecological validity. With the introduction of EgoLive, we aim to provide the research community with a scalable, high-quality dataset that accelerates breakthroughs in generalizable robotic models and facilitates the real-world deployment of robot systems.