Learning Dexterous Grasping from Sparse Taxonomy Guidance

arXiv cs.RO / 4/7/2026

📰 NewsSignals & Early TrendsModels & Research

Key Points

  • The paper introduces GRIT, a two-stage framework for dexterous grasping that uses sparse taxonomy guidance rather than dense grasp/contact supervision.
  • GRIT first predicts a taxonomy-based grasp specification from scene and task context, then generates continuous multi-finger motions conditioned on that sparse grasp structure.
  • The authors find that different grasp taxonomies work better for different object geometries, and they leverage this relationship to improve generalization.
  • On benchmark experiments, GRIT reports an overall success rate of 87.9% and improved performance on novel objects versus baseline methods.
  • Real-world tests indicate the approach is controllable, allowing grasp strategies to be adjusted via high-level taxonomy selection aligned with object geometry and task intent.

Abstract

Dexterous manipulation requires planning a grasp configuration suited to the object and task, which is then executed through coordinated multi-finger control. However, specifying grasp plans with dense pose or contact targets for every object and task is impractical. Meanwhile, end-to-end reinforcement learning from task rewards alone lacks controllability, making it difficult for users to intervene when failures occur. To this end, we present GRIT, a two-stage framework that learns dexterous control from sparse taxonomy guidance. GRIT first predicts a taxonomy-based grasp specification from the scene and task context. Conditioned on this sparse command, a policy generates continuous finger motions that accomplish the task while preserving the intended grasp structure. Our result shows that certain grasp taxonomies are more effective for specific object geometries. By leveraging this relationship, GRIT improves generalization to novel objects over baselines and achieves an overall success rate of 87.9%. Moreover, real-world experiments demonstrate controllability, enabling grasp strategies to be adjusted through high-level taxonomy selection based on object geometry and task intent.