SpikeGrasp: A Benchmark for 6-DoF Grasp Pose Detection from Stereo Spike Streams

arXiv cs.RO / 3/23/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • SpikeGrasp introduces a neuro-inspired end-to-end framework for 6-DoF grasp pose detection using raw asynchronous events from stereo spike cameras, avoiding explicit 3D point clouds.
  • It uses a recurrent spiking neural network to iteratively refine grasp hypotheses, mirroring a biological visuomotor pathway.
  • The approach is validated on a large-scale synthetic benchmark dataset to assess performance.
  • Experiments indicate SpikeGrasp outperforms traditional point-cloud baselines, especially in cluttered or textureless scenes, highlighting data efficiency and potential for real-time, dynamic robotics.

Abstract

Most robotic grasping systems rely on converting sensor data into explicit 3D point clouds, which is a computational step not found in biological intelligence. This paper explores a fundamentally different, neuro-inspired paradigm for 6-DoF grasp detection. We introduce SpikeGrasp, a framework that mimics the biological visuomotor pathway, processing raw, asynchronous events from stereo spike cameras, similarly to retinas, to directly infer grasp poses. Our model fuses these stereo spike streams and uses a recurrent spiking neural network, analogous to high-level visual processing, to iteratively refine grasp hypotheses without ever reconstructing a point cloud. To validate this approach, we built a large-scale synthetic benchmark dataset. Experiments show that SpikeGrasp surpasses traditional point-cloud-based baselines, especially in cluttered and textureless scenes, and demonstrates remarkable data efficiency. By establishing the viability of this end-to-end, neuro-inspired approach, SpikeGrasp paves the way for future systems capable of the fluid and efficient manipulation seen in nature, particularly for dynamic objects.