AI Navigate

dTRPO: Trajectory Reduction in Policy Optimization of Diffusion Large Language Models

arXiv cs.AI / 3/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • dTRPO introduces trajectory reduction techniques to cut the cost of trajectory probability calculation in diffusion LLM policy optimization, enabling scalable offline training.
  • It proves that under reference policy regularization, the probability ratio of newly unmasked tokens is an unbiased estimate of the ratio for intermediate diffusion states, and that the full trajectory probability can be estimated with a single forward pass of a re-masked final state.
  • By integrating these results into a policy optimization objective, dTRPO achieves gains on 7B dLLMs across STEM tasks (up to 9.6%), coding tasks (up to 4.3%), and instruction-following tasks (up to 3.0%).
  • It also delivers training efficiency via offline, single-forward evaluation and improved generation efficiency through high-quality outputs.

Abstract

Diffusion Large Language Models (dLLMs) introduce a new paradigm for language generation, which in turn presents new challenges for aligning them with human preferences. In this work, we aim to improve the policy optimization for dLLMs by reducing the cost of the trajectory probability calculation, thereby enabling scaled-up offline policy training. We prove that: (i) under reference policy regularization, the probability ratio of the newly unmasked tokens is an unbiased estimate of that of intermediate diffusion states, and (ii) the probability of the full trajectory can be effectively estimated with a single forward pass of a re-masked final state. By integrating these two trajectory reduction strategies into a policy optimization objective, we propose Trajectory Reduction Policy Optimization (dTRPO). We evaluate dTRPO on 7B dLLMs across instruction-following and reasoning benchmarks. Results show that it substantially improves the core performance of state-of-the-art dLLMs, achieving gains of up to 9.6% on STEM tasks, up to 4.3% on coding tasks, and up to 3.0% on instruction-following tasks. Moreover, dTRPO exhibits strong training efficiency due to its offline, single-forward nature, and achieves improved generation efficiency through high-quality outputs.