FlashCap: Millisecond-Accurate Human Motion Capture via Flashing LEDs and Event-Based Vision

arXiv cs.CV / 3/23/2026

💬 OpinionSignals & Early TrendsModels & Research

Key Points

  • The paper introduces FlashCap, a millisecond-accurate motion capture system that uses flashing LEDs and event-based vision to enable precise motion timing in human pose estimation.
  • FlashCap enables the FlashMotion dataset, a millisecond-resolution multimodal collection (event data, RGB, LiDAR, and IMU) designed to close the high-temporal-resolution data gap for PMT.
  • The study proposes ResPose, a residual-pose learning baseline that fuses events and RGBs and reduces pose estimation error by about 40%.
  • The authors will share the dataset and code with the community to foster new research opportunities in high-temporal-resolution HPE and PMT.

Abstract

Precise motion timing (PMT) is crucial for swift motion analysis. A millisecond difference may determine victory or defeat in sports competitions. Despite substantial progress in human pose estimation (HPE), PMT remains largely overlooked by the HPE community due to the limited availability of high-temporal-resolution labeled datasets. Today, PMT is achieved using high-speed RGB cameras in specialized scenarios such as the Olympic Games; however, their high costs, light sensitivity, bandwidth, and computational complexity limit their feasibility for daily use. We developed FlashCap, the first flashing LED-based MoCap system for PMT. With FlashCap, we collect a millisecond-resolution human motion dataset, FlashMotion, comprising the event, RGB, LiDAR, and IMU modalities, and demonstrate its high quality through rigorous validation. To evaluate the merits of FlashMotion, we perform two tasks: precise motion timing and high-temporal-resolution HPE. For these tasks, we propose ResPose, a simple yet effective baseline that learns residual poses based on events and RGBs. Experimental results show that ResPose reduces pose estimation errors by ~40% and achieves millisecond-level timing accuracy, enabling new research opportunities. The dataset and code will be shared with the community.