SurfaceXR: Fusing Smartwatch IMUs and Egocentric Hand Pose for Seamless Surface Interactions

arXiv cs.CV / 3/23/2026

💬 OpinionModels & Research

Key Points

  • SurfaceXR fuses headset-based hand tracking with smartwatch IMU data to enable robust surface-based inputs in XR, addressing fatigue and imprecision of mid-air gestures.
  • The approach leverages complementary modalities: 3D hand pose from vision and high-frequency motion from IMUs to improve accuracy on everyday surfaces.
  • A 21-participant study demonstrated improved touch tracking and 8-class gesture recognition compared with single-modality methods.
  • The work aims to solve egocentric vision hand-tracking challenges and unreliable surface plane estimation, offering a more comfortable and reliable interaction method for XR users.

Abstract

Mid-air gestures in Extended Reality (XR) often cause fatigue and imprecision. Surface-based interactions offer improved accuracy and comfort, but current egocentric vision methods struggle due to hand tracking challenges and unreliable surface plane estimation. We introduce SurfaceXR, a sensor fusion approach combining headset-based hand tracking with smartwatch IMU data to enable robust inputs on everyday surfaces. Our insight is that these modalities are complementary: hand tracking provides 3D positional data while IMUs capture high-frequency motion. A 21-participant study validates SurfaceXR's effectiveness for touch tracking and 8-class gesture recognition, demonstrating significant improvements over single-modality approaches.
広告