Conformalized Percentile Interval: Finite Sample Validity and Improved Conditional Performance

arXiv stat.ML / 5/6/2026

📰 NewsModels & Research

Key Points

  • The paper introduces a conformal-style method that builds finite-sample-adjusted percentile prediction intervals using PIT (probability integral transform) values from a neural-network-estimated conditional CDF.
  • It argues that calibrating in PIT space reduces feature-dependent miscoverage when the CDF estimator is accurate, improving conditional validity and calibration.
  • The approach is designed to remain robust even if the conditional CDF is imperfect, by leveraging the empirical PIT distribution for percentile calibration.
  • The authors provide theoretical guarantees, including finite-sample marginal coverage and asymptotic conditional coverage under mild consistency assumptions.
  • Experiments on synthetic and real-world benchmarks show improved conditional calibration and substantially shorter intervals compared with existing methods.

Abstract

Conformal prediction provides distribution-free predictive intervals with finite-sample marginal coverage. However, achieving conditional validity and interval efficiency (in terms of short interval length) remains challenging, particularly in complex settings with heteroskedasticity, skewed responses, or estimation errors. We propose a conformal-style calibration method for responses obtained by the probability integral transform (PIT) of the conditional cumulative distribution function (CDF) estimated via neural networks to construct a finite-sample-adjusted percentile interval with the shortest length determined by the estimated conditional CDF. Calibrating in PIT space is effective because PIT values are asymptotically feature-independent when the CDF estimator is accurate, which mitigates feature-dependent miscoverage and improves conditional calibration. On the other hand, our percentile calibration adapts to the empirical PIT distribution, which is robust against a possibly imperfect estimation of the conditional CDF. We prove the finite-sample marginal coverage property of the proposed method and show its asymptotic conditional coverage under mild consistency conditions. Experiments on diverse synthetic and real-world benchmarks demonstrate better conditional calibration and substantially shorter intervals than existing methods.