DORA: A Scalable Asynchronous Reinforcement Learning System for Language Model Training

arXiv cs.LG / 4/30/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper introduces DORA, a system for scalable asynchronous reinforcement learning (RL) during language model post-training, targeting bottlenecks caused by skewed, long-tailed generation trajectories.
  • DORA addresses convergence-critical requirements for asynchronous RL by enforcing intra-trajectory policy consistency, data integrity, and bounded staleness, which prior approaches either miss or handle only partially.
  • The core method, multi-version streaming rollout, keeps multiple policy versions active concurrently to eliminate training “bubbles” while preserving the algorithmic constraints needed for convergence.
  • Experiments show throughput gains of 2–3× over state-of-the-art on open benchmarks without harming convergence, and 2–4× speedups versus synchronous training in large industrial settings using tens of thousands of accelerators.
  • The work releases open-source models (LongCat-Flash-Thinking) that achieve competitive results on complex reasoning benchmarks, comparable to many advanced LLMs.

Abstract

Reinforcement learning (RL) has become a critical paradigm for LLM post-training, yet the rollout phase -- accounting for 50--80% of total step time -- is bottlenecked by skewed generation: long-tailed trajectories indispensable for model performance block the entire training pipeline. Asynchronous training offers a natural remedy by overlapping generation with training, but introduces a fundamental tension between efficiency and algorithmic correctness. We identify three constraints in asynchronous training to preserve convergence: intra-trajectory policy consistency, data integrity, and bounded staleness. Existing approaches fail to intrinsically address the long-tailed trajectory problem, which is further exacerbated by the imbalance characteristic of Mix-of-Experts models, or deviate from the standard RL training formulation, thereby hindering model convergence. Therefore, we propose DORA (Dynamic ORchestration for Asynchronous Rollout), which addresses this challenge through algorithm-system co-design. DORA introduces multi-version streaming rollout, a novel asynchronous paradigm that maintains multiple policy versions concurrently -- simultaneously achieving full bubble elimination without compromising algorithmic constraints. Experimental results demonstrate that our DORA system achieves substantial improvements in throughput -- up to 2--3 times higher than state-of-the-art systems on open-source benchmarks -- without compromising convergence. Furthermore, in large-scale industrial applications with tens of thousands of accelerators, DORA accelerates RL training by 2--4 times compared to synchronous training across various scenarios. The resultant open-source models, LongCat-Flash-Thinking, exhibit competitive performance on complex reasoning benchmarks, matching the capability of most advanced LLMs.