UniDex: A Robot Foundation Suite for Universal Dexterous Hand Control from Egocentric Human Videos
arXiv cs.RO / 2026/3/24
📰 ニュースSignals & Early TrendsIdeas & Deep AnalysisModels & Research
要点
- UniDex proposes a “robot foundation suite” for universal dexterous hand control by combining a large robot-centric dataset, a unified vision-language-action (VLA) policy, and a human-data capture setup.
- The approach builds UniDex-Dataset with 50K+ robot trajectories across eight different dexterous hands (6–24 DoFs) by retargeting egocentric human videos via a human-in-the-loop procedure that preserves plausible hand-object contacts.
- UniDex introduces FAAS (Function-Actuator-Aligned Space), an action-space mapping that aligns functionally similar actuators across different hand embodiments to enable cross-hand transfer.
- A pretrained UniDex-VLA policy is trained on the robot-centric dataset and then fine-tuned with task demonstrations, achieving strong performance on tool-use tasks (81% average task progress) and notable zero-shot cross-hand generalization.
- UniDex-Cap provides a portable RGB-D + hand-pose capture pipeline to convert human data into robot-executable trajectories, aiming to reduce dependence on expensive robot teleoperation for co-training.
