MolmoB0T: Large-Scale Simulation Enables Zero-Shot Manipulation

arXiv cs.RO / 3/27/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper challenges the common belief that sim-to-real for robotic manipulation requires real-world data collection or task-specific fine-tuning, showing strong zero-shot transfer using only large-scale, diverse synthetic simulation training.
  • It introduces MolmoBot-Engine, an open-source pipeline for procedural generation of training data across different robots, tasks, and simulated environments in MolmoSpaces.
  • It releases MolmoBot-Data, a dataset containing 1.8M expert trajectories for articulated object manipulation and pick-and-place tasks, supporting training and benchmarking.
  • Three policy variants are trained and compared, including a Molmo2-based multi-frame vision-language model (MolmoBot), a pi0-replicating baseline (MolmoBot-Pi0), and a lightweight edge-oriented policy (MolmoBot-SPOC).
  • Evaluations on Franka FR3 and Rainbow Robotics RB-Y1 show no real-world fine-tuning yields effective zero-shot manipulation, with tabletop pick-and-place reaching 79.2% success versus 39.2% for pi0.5 across four settings.

Abstract

A prevailing view in robot learning is that simulation alone is not enough; effective sim-to-real transfer is widely believed to require at least some real-world data collection or task-specific fine-tuning to bridge the gap between simulated and physical environments. We challenge that assumption. With sufficiently large-scale and diverse simulated synthetic training data, we show that zero-shot transfer to the real world is not only possible, but effective for both static and mobile manipulation. We introduce MolmoBot-Engine, a fully open-source pipeline for procedural data generation across robots, tasks, and diverse simulated environments in MolmoSpaces. With it, we release MolmoBot-Data, a dataset of 1.8 million expert trajectories for articulated object manipulation and pick-and-place tasks. We train three policy classes: MolmoBot, a Molmo2-based multi-frame vision-language model with a flow-matching action head; MolmoBot-Pi0, which replicates the \pi_0 architecture to enable direct comparison; and MolmoBot-SPOC, a lightweight policy suitable for edge deployment and amenable to RL fine-tuning. We evaluate on two robotic platforms: the Franka FR3 for tabletop manipulation tasks and the Rainbow Robotics RB-Y1 mobile manipulator for door opening, drawer manipulation, cabinet interaction, and mobile pick-and-place. Without any real-world fine-tuning, our policies achieve zero-shot transfer to unseen objects and environments. On tabletop pick-and-place, MolmoBot achieves a success rate of 79.2% in real world evaluations across 4 settings, outperforming \pi_{0.5} at 39.2%. Our results demonstrate that procedural environment generation combined with diverse articulated assets can produce robust manipulation policies that generalize broadly to the real world. Technical website: https://allenai.github.io/MolmoBot
広告