Lightweight Learning from Actuation-Space Demonstrations via Flow Matching for Whole-Body Soft Robotic Grasping

arXiv cs.RO / 4/6/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses uncertainty and contact-rich challenges in robotic grasping by leveraging soft robots’ passive compliance and underactuated whole-body mechanics for more adaptable behavior.
  • It proposes a “lightweight actuation-space learning” framework that learns distributional control representations for whole-body soft robotic grasping from deterministic demonstrations using a flow-matching model (Rectified Flow).
  • The approach avoids dense sensing and heavy feedback control loops, aiming to shift control intelligence from complex central controllers to the robot’s embodied mechanics.
  • With only 30 demonstrations (under 8% of the workspace), the method reportedly achieves a 97.5% grasp success rate across the full workspace.
  • The learned policy also generalizes to object size changes (±33%) and remains stable under direct dynamic-response adjustments by scaling execution time (20% to 200%).

Abstract

Robotic grasping under uncertainty remains a fundamental challenge due to its uncertain and contact-rich nature. Traditional rigid robotic hands, with limited degrees of freedom and compliance, rely on complex model-based and heavy feedback controllers to manage such interactions. Soft robots, by contrast, exhibit embodied mechanical intelligence: their underactuated structures and passive flexibility of their whole body, naturally accommodate uncertain contacts and enable adaptive behaviors. To harness this capability, we propose a lightweight actuation-space learning framework that infers distributional control representations for whole-body soft robotic grasping, directly from deterministic demonstrations using a flow matching model (Rectified Flow),without requiring dense sensing or heavy control loops. Using only 30 demonstrations (less than 8% of the reachable workspace), the learned policy achieves a 97.5% grasp success rate across the whole workspace, generalizes to grasped-object size variations of +-33%, and maintains stable performance when the robot's dynamic response is directly adjusted by scaling the execution time from 20% to 200%. These results demonstrate that actuation-space learning, by leveraging its passive redundant DOFs and flexibility, converts the body's mechanics into functional control intelligence and substantially reduces the burden on central controllers for this uncertain-rich task.