Two-sample comparison through additive tree models for density ratios
arXiv stat.ML / 4/23/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper tackles two-sample comparison by estimating the density ratio between two distributions from i.i.d. samples, using additive tree models.
- It introduces a new training objective called “balancing loss,” which enables tree models to be optimized with supervised-learning algorithms such as forward-stagewise optimization and gradient boosting.
- The balancing loss is shown to relate to exponential-family kernels and can be used as a pseudo-likelihood, enabling generalized Bayesian inference via backfitting samplers for Bayesian additive regression trees (BART).
- The authors provide uncertainty quantification for the estimated density ratio, and they link the balancing loss to binary classification losses and to variational forms of f-divergences (notably squared Hellinger distance).
- Experiments indicate improved accuracy and computational efficiency, and the method is demonstrated on evaluating generative models for microbiome compositional data.
Related Articles

Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to

Trajectory Forecasts in Unknown Environments Conditioned on Grid-Based Plans
Dev.to

Why use an AI gateway at all?
Dev.to

OpenAI Just Named It Workspace Agents. We Open-Sourced Our Lark Version Six Months Ago
Dev.to

GPT Image 2 Subject-Lock Editing: A Practical Guide to input_fidelity
Dev.to