Genie Sim 3.0 : A High-Fidelity Comprehensive Simulation Platform for Humanoid Robot

arXiv cs.RO / 4/29/2026

💬 OpinionDeveloper Stack & InfrastructureSignals & Early TrendsModels & Research

Key Points

  • Genie Sim 3.0 introduces a unified, high-fidelity simulation platform aimed at improving the training and evaluation of generalizable robot manipulation learning models.
  • The system includes a Genie Sim Generator that uses an LLM to create high-fidelity scenes from natural-language instructions, enabling rapid and multi-dimensional generalization across diverse environments.
  • The article presents a new benchmark that pioneers using LLMs for automated evaluation by generating many evaluation scenarios and using a vision-language model (VLM) to run an assessment pipeline.
  • An open-source dataset is released with 10,000+ hours of synthetic data spanning 200+ tasks, and experiments show robust zero-shot sim-to-real transfer under controlled conditions.
  • The authors position synthetic data as an effective substitute for real-world data for scalable policy training, while providing code and dataset links for replication and reuse.

Abstract

The development of robust and generalizable robot learning models is critically contingent upon the availability of large-scale, diverse training data and reliable evaluation benchmarks. Collecting data in the physical world poses prohibitive costs and scalability challenges, and prevailing simulation benchmarks frequently suffer from fragmentation, narrow scope, or insufficient fidelity to enable effective sim-to-real transfer. To address these challenges, we introduce Genie Sim 3.0, a unified simulation platform for robotic manipulation. We present Genie Sim Generator, a large language model (LLM)-powered tool that constructs high-fidelity scenes from natural language instructions. Its principal strength resides in rapid and multi-dimensional generalization, facilitating the synthesis of diverse environments to support scalable data collection and robust policy evaluation. We introduce the first benchmark that pioneers the application of LLM for automated evaluation. It leverages LLM to mass-generate evaluation scenarios and employs Vision-Language Model (VLM) to establish an automated assessment pipeline. We also release an open-source dataset comprising more than 10,000 hours of synthetic data across over 200 tasks. Through systematic experimentation, we validate the robust zero-shot sim-to-real transfer capability of our open-source dataset, demonstrating that synthetic data can server as an effective substitute for real-world data under controlled conditions for scalable policy training. For code and dataset details, please refer to: https://github.com/AgibotTech/genie_sim.