D3-Gym: Constructing Real-World Verifiable Environments for Data-Driven Discovery

arXiv cs.AI / 5/1/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • D3-Gym is introduced to address a gap in scientific data-driven discovery by providing verifiable environments that represent real-world scientific tasks.
  • The dataset includes 565 tasks from 239 real scientific repositories across four disciplines, with each task packaged with instructions, an executable environment, input data/preview artifacts, reference code, and an automatically generated evaluation script.
  • The authors report strong verification quality: the synthesized evaluation scripts reach 87.5% agreement with human-labeled gold standards and show solid alignment with domain-specific evaluation logic.
  • Training on D3-Gym trajectories reportedly improves multiple Qwen3 model variants on ScienceAgentBench, including a 7.8-point boost for Qwen3-32B and a reduced gap versus strong proprietary models.
  • All environments, workflows, trajectories, and models are released publicly on GitHub for reuse and further research.

Abstract

Despite recent progress in language models and agents for scientific data-driven discovery, further advancing their capabilities is held back by the absence of verifiable environments representing real-world scientific tasks.To fill this gap, we introduce D3-Gym, the first automatically constructed dataset with verifiable environments for scientific Data-Driven Discovery. D3-Gym comprises (1) 565 tasks sourced from 239 real scientific repositories across four disciplines where (2) each task is equipped with a natural language instruction, an executable environment with pre-installed dependencies, input dataset and artifact previews, a reference code solution, and an automatically synthesized evaluation script. Rigorous evaluation of the quality of the verification signal in D3-Gym confirms that our evaluation scripts achieve 87.5% agreement with human-annotated gold standards and strong alignment in domain-specific evaluation logic, showing their scientific soundness. Further, training on trajectories sampled from D3-Gym yields consistent and substantial gains across Qwen3 models of varying sizes on ScienceAgentBench, boosting Qwen3-32B by 7.8 absolute points and substantially shrinking the gap with strong proprietary models. All D3-Gym artifacts (environments, creation workflow, trajectories, and models) can be found at https://github.com/OSU-NLP-Group/D3-Gym.