EgoMAGIC- An Egocentric Video Field Medicine Dataset for Training Perception Algorithms
arXiv cs.AI / 4/27/2026
💬 OpinionSignals & Early TrendsTools & Practical UsageModels & Research
Key Points
- The paper introduces EgoMAGIC, an egocentric medical activity video dataset collected under DARPA’s Perceptually-enabled Task Guidance (PTG) program to support AR-assisted virtual task guidance.
- EgoMAGIC includes 3,355 videos covering 50 medical tasks, with at least 50 labeled videos per task, and most videos recorded with a head-mounted stereo camera and audio.
- The authors also release medical training data and a focused action detection challenge covering eight medical tasks, lowering the barrier for research and experimentation.
- Using this dataset, the team trained 40 YOLO models with 1.95 million labels to detect 124 medical objects, and reports baseline action-detection results with a best average mAP of 0.526.
- Beyond action detection, the dataset is positioned as useful for other computer-vision tasks such as action recognition, object identification/detection, and error detection.




