DARE: Diffusion Large Language Models Alignment and Reinforcement Executor

arXiv cs.CL / 4/7/2026

📰 NewsDeveloper Stack & InfrastructureSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces DARE, an open framework designed to unify post-training pipelines for diffusion large language models (dLLMs), including supervised fine-tuning, parameter-efficient fine-tuning, preference optimization, and dLLM-specific reinforcement learning.
  • It targets the fragmented research ecosystem for dLLM alignment and RL, where objectives, rollout implementations, and evaluation code are often released as paper-specific artifacts, making reproduction and fair comparison difficult.
  • DARE is built on top of verl and OpenCompass, providing a shared execution stack that supports both masked and block diffusion language models.
  • The framework covers multiple representative dLLM model families (LLaDA, Dream, SDAR, and LLaDA2.x) and aims to deliver reproducible benchmark evaluation plus practical acceleration.
  • The authors present extensive empirical results positioning DARE as a reusable research “substrate” for developing, comparing, and deploying current and emerging dLLM post-training methods.

Abstract

Diffusion large language models (dLLMs) are emerging as a compelling alternative to dominant autoregressive models, replacing strictly sequential token generation with iterative denoising and parallel generation dynamics. However, their open-source ecosystem remains fragmented across model families and, in particular, across post-training pipelines, where reinforcement learning objectives, rollout implementations and evaluation scripts are often released as paper-specific codebases. This fragmentation slows research iteration, raises the engineering burden of reproduction, and makes fair comparison across algorithms difficult. We present \textbf{DARE} (\textbf{d}LLMs \textbf{A}lignment and \textbf{R}einforcement \textbf{E}xecutor), an open framework for post-training and evaluating dLLMs. Built on top of verl~\cite{sheng2024hybridflow} and OpenCompass~\cite{2023opencompass}, DARE unifies supervised fine-tuning, parameter-efficient fine-tuning, preference optimization, and dLLM-specific reinforcement learning under a shared execution stack for both masked and block diffusion language models. Across representative model families including LLaDA, Dream, SDAR, and LLaDA2.x, DARE provides broad algorithmic coverage, reproducible benchmark evaluation, and practical acceleration. Extensive empirical results position that DARE serves as a reusable research substrate for developing, comparing, and deploying post-training methods for current and emerging dLLMs.