A Multimodal Data Collection Framework for Dialogue-Driven Assistive Robotics to Clarify Ambiguities: A Wizard-of-Oz Pilot Study
arXiv cs.RO / 4/17/2026
📰 News
Key Points
- The paper addresses a key limitation in assistive robotics: current interfaces and datasets do not adequately capture multimodal, dialogue-driven ambiguity in natural human-robot interaction.
- It proposes a multimodal data collection framework using a dialogue-based protocol and a two-room Wizard-of-Oz setup to simulate robot autonomy while encouraging natural user behavior.
- The system records five synchronized modalities—RGB-D video, conversational audio, IMU signals, end-effector Cartesian pose, and whole-body joint states—across five wheelchair/robot-arm assistance tasks.
- A pilot dataset of 53 trials from five participants was collected and evaluated using motion-smoothness analysis and user feedback, showing the approach can represent diverse ambiguity types.
- The authors argue the framework is suitable for scaling to larger datasets to support training, benchmarking, and evaluation of ambiguity-aware assistive control.
- categories: [