Trees to Flows and Back: Unifying Decision Trees and Diffusion Models
arXiv cs.AI / 5/4/2026
💬 OpinionIdeas & Deep AnalysisTools & Practical UsageModels & Research
Key Points
- The paper establishes a mathematical correspondence between hierarchical decision trees and diffusion processes under suitable limiting regimes, showing they can be unified despite different data/model structures.
- It identifies a shared optimization principle called Global Trajectory Score Matching (GTSM), and argues that (idealized) gradient boosting is asymptotically optimal for it.
- The authors demonstrate the ideas with two practical implementations: treeflow for faster, high-fidelity tabular data generation and dsmtree for distilling decision-tree logic into neural networks.
- In experiments, treeflow reportedly delivers competitive generation quality with higher fidelity and about a 2× computational speedup, while dsmtree matches teacher performance within ~2% on many benchmarks.
- Overall, the work provides both a theoretical bridge and concrete methods that connect discrete decision logic to modern diffusion-based generative modeling.



