OmniRobotHome: A Multi-Camera Platform for Real-Time Multiadic Human-Robot Interaction

arXiv cs.RO / 5/1/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper argues that real home settings require “multiadic” human-robot collaboration (multiple humans and robots acting concurrently), but this has been hard to study due to persistent occlusion and rapidly changing states.
  • It introduces OmniRobotHome, a room-scale platform that unifies real-time, occlusion-robust 3D perception with coordinated multi-robot actuation in a shared world coordinate frame.
  • The system uses 48 hardware-synchronized RGB cameras to enable markerless tracking of multiple people and objects, aligned in time with two Franka robotic arms that respond to live scene state.
  • The platform also supports long-horizon human behavior modeling by maintaining continuous capture within a consistent frame, which can accumulate trajectories over time.
  • The authors demonstrate measurable improvements for two goals—safety in shared human-robot environments and human-anticipatory robotic assistance—coming from both real-time perception and accumulated behavior memory.

Abstract

Human-robot collaboration has been studied primarily in dyadic or sequential settings. However, real homes require multiadic collaboration, where multiple humans and robots share a workspace, acting concurrently on interleaved subtasks with tight spatial and temporal coupling. This regime remains underexplored because close-proximity interaction between humans, robots, and objects creates persistent occlusion and rapid state changes, making reliable real-time 3D tracking the central bottleneck. No existing platform provides the real-time, occlusion-robust, room-scale perception needed to make this regime experimentally tractable. We present OmniRobotHome, the first room-scale residential platform that unifies wide-area real-time 3D human and object perception with coordinated multi-robot actuation in a shared world frame. The system instruments a natural home environment with 48 hardware-synchronized RGB cameras for markerless, occlusion-robust tracking of multiple humans and objects, temporally aligned with two Franka arms that act on live scene state. Continuous capture within this consistent frame further supports long-horizon human behavior modeling from accumulated trajectories. The platform makes the multiadic collaboration regime experimentally tractable. We focus on two central problems: safety in shared human-robot environments and human-anticipatory robotic assistance, and show that real-time perception and accumulated behavior memory each yield measurable gains in both.