Emergent Dexterity via Diverse Resets and Large-Scale Reinforcement Learning

arXiv cs.RO / 3/25/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper argues that existing sim-to-real dexterous manipulation reinforcement learning methods are brittle, task-specific, and do not scale with compute due to performance saturating from limited state-space coverage.
  • It introduces OmniReset, a framework that uses diverse simulator resets to expose an on-policy RL agent to a wide range of robot-object interactions without relying on curricula, reward redesign per task, or human demonstrations.
  • OmniReset keeps a single reward function and fixed algorithm hyperparameters across tasks, aiming to remove the heavy per-task engineering burden common in prior approaches.
  • Experiments show improved scaling to long-horizon, contact-rich manipulation tasks and robust policies over a wider range of initial conditions than baseline methods.
  • The authors distill OmniReset-trained policies into visuomotor behaviors that demonstrate higher real-world zero-shot transfer success rates and robust “retrying” behavior.

Abstract

Reinforcement learning in massively parallel physics simulations has driven major progress in sim-to-real robot learning. However, current approaches remain brittle and task-specific, relying on extensive per-task engineering to design rewards, curricula, and demonstrations. Even with this engineering, they often fail on long-horizon, contact-rich manipulation tasks and do not meaningfully scale with compute, as performance quickly saturates when training revisits the same narrow regions of state space. We introduce OmniReset, a simple and scalable framework that enables on-policy reinforcement learning to robustly solve a broad class of dexterous manipulation tasks using a single reward function, fixed algorithm hyperparameters, no curricula, and no human demonstrations. Our key insight is that long-horizon exploration can be dramatically simplified by using simulator resets to systematically expose the RL algorithm to the diverse set of robot-object interactions which underlie dexterous manipulation. OmniReset programmatically generates such resets with minimal human input, converting additional compute directly into broader behavioral coverage and continued performance gains. We show that OmniReset gracefully scales to long-horizon dexterous manipulation tasks beyond the capabilities of existing approaches and is able to learn robust policies over significantly wider ranges of initial conditions than baselines. Finally, we distill OmniReset into visuomotor policies which display robust retrying behavior and substantially higher success rates than baselines when transferred to the real world zero-shot. Project webpage: https://omnireset.github.io