UMI-Underwater: Learning Underwater Manipulation without Underwater Teleoperation
arXiv cs.RO / 3/31/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The UMI-Underwater system addresses challenging underwater robotic grasping by combining autonomous self-supervised data collection with learning strategies that reduce the need for diverse underwater demonstrations.
- It transfers grasp knowledge from on-land human handheld demos to underwater using a depth-based affordance representation designed to bridge the on-land-to-underwater domain gap and remain robust to lighting and color shifts.
- An affordance model trained on on-land data is deployed underwater in a zero-shot manner via geometric alignment before training a diffusion-based control policy conditioned on affordances.
- Pool experiments show improved grasp performance and robustness to background changes, along with better generalization to objects only seen in on-land data compared with RGB-only baselines.
- The work provides code, videos, and additional results publicly via its project website.
Related Articles

Black Hat Asia
AI Business
[D] How does distributed proof of work computing handle the coordination needs of neural network training?
Reddit r/MachineLearning

Claude Code's Entire Source Code Was Just Leaked via npm Source Maps — Here's What's Inside
Dev.to

BYOK is not just a pricing model: why it changes AI product trust
Dev.to

AI Citation Registries and Identity Persistence Across Records
Dev.to