AI Navigate

Learning Permutation Distributions via Reflected Diffusion on Ranks

arXiv cs.LG / 3/19/2026

📰 NewsModels & Research

Key Points

  • It proposes Soft-Rank Diffusion, a discrete diffusion framework that models permutations by lifting them to a continuous soft-rank representation to enable smoother, more tractable trajectories.
  • The forward process replaces shuffle-based corruption with a structured soft-rank noise, addressing challenges from factorial growth in S_n.
  • The reverse process introduces contextualized generalized Plackett-Luce (cGPL) denoisers to better capture sequential decision structures in permutations.
  • Experiments on sorting and combinatorial optimization show Soft-Rank Diffusion outperforming prior diffusion baselines, with strong gains in long-sequence settings.
  • The approach promises improved permutation modeling for tasks like ranking and scheduling, where scalable, accurate distributions over permutations are valuable.

Abstract

The finite symmetric group S_n provides a natural domain for permutations, yet learning probability distributions on S_n is challenging due to its factorially growing size and discrete, non-Euclidean structure. Recent permutation diffusion methods define forward noising via shuffle-based random walks (e.g., riffle shuffles) and learn reverse transitions with Plackett-Luce (PL) variants, but the resulting trajectories can be abrupt and increasingly hard to denoise as n grows. We propose Soft-Rank Diffusion, a discrete diffusion framework that replaces shuffle-based corruption with a structured soft-rank forward process: we lift permutations to a continuous latent representation of order by relaxing discrete ranks into soft ranks, yielding smoother and more tractable trajectories. For the reverse process, we introduce contextualized generalized Plackett-Luce (cGPL) denoisers that generalize prior PL-style parameterizations and improve expressivity for sequential decision structures. Experiments on sorting and combinatorial optimization benchmarks show that Soft-Rank Diffusion consistently outperforms prior diffusion baselines, with particularly strong gains in long-sequence and intrinsically sequential settings.