Tube Diffusion Policy: Reactive Visual-Tactile Policy Learning for Contact-rich Manipulation
arXiv cs.RO / 4/28/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces Tube Diffusion Policy (TDP), a reactive visual-tactile imitation learning framework designed for contact-rich manipulation under uncertainty and disturbances.
- TDP combines diffusion-based imitation with tube-based feedback control, learning an observation-conditioned feedback flow around nominal action chunks to form an “action tube” for rapid corrections during execution.
- Experiments on the Push-T benchmark plus three additional visual-tactile dexterous manipulation tasks show that TDP outperforms existing imitation learning baselines consistently.
- Real-world tests confirm TDP’s robustness in handling contact uncertainty and external disturbances, and its step-wise correction reduces the number of denoising steps for real-time high-frequency control.
- The proposed tube-based mechanism addresses a key limitation of action-chunking approaches by enabling reaction to unforeseen observations during execution.
Related Articles

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Same Agent, Different Risk | How Microsoft 365 Copilot Grounding Changes the Security Model | Rahsi Framework™
Dev.to

Claude Haiku for Low-Cost AI Inference: Patterns from a Horse Racing Prediction System
Dev.to

How We Built an Ambient AI Clinical Documentation Pipeline (and Saved Doctors 8+ Hours a Week)
Dev.to

🦀 PicoClaw Deep Dive — A Field Guide to Building an Ultra-Light AI Agent in Go 🐹
Dev.to