HiSync: Spatio-Temporally Aligning Hand Motion from Wearable IMU and On-Robot Camera for Command Source Identification in Long-Range HRI
arXiv cs.RO / 3/26/2026
💬 OpinionIdeas & Deep AnalysisTools & Practical UsageModels & Research
Key Points
- The paper proposes HiSync, an optical-inertial fusion framework that aligns a robot-mounted camera’s optical flow with a hand-worn IMU to identify the command source in long-range, multi-user HRI.
- HiSync learns frequency-domain features from both modalities, denoises IMU signals with CSINet, and uses temporal alignment plus distance-aware multi-window fusion to match subtle natural gestures.
- The authors collect a user-defined gesture set (N=12) and a multimodal command gesture dataset (N=38) for long-range multi-user scenarios, targeting the ambiguity created by distance and multiple users.
- In three-person scenes up to 34 meters, HiSync reports 92.32% CSI accuracy and claims a 48.44% improvement over prior state of the art, and it is validated through real-robot deployment.
- The work is positioned as a practical HRI primitive and provides design guidance, with code released on GitHub for reproducibility and further development.
Related Articles
Regulating Prompt Markets: Securities Law, Intellectual Property, and the Trading of Prompt Assets
Dev.to
Mercor competitor Deccan AI raises $25M, sources experts from India
Dev.to
How We Got Local MCP Servers Working in Claude Cowork (The Missing Guide)
Dev.to
How Should Students Document AI Usage in Academic Work?
Dev.to
I built a PWA fitness tracker with AI that supports 86 sports — as a solo developer
Dev.to