Online monotone density estimation and log-optimal calibration

arXiv stat.ML / 3/31/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies online monotone density estimation where estimators are built in a predictable way from sequential observations under a monotonicity assumption.
  • It introduces two methods—an online version of the Grenander estimator and an expert aggregation estimator based on exponential-weighting ideas from online learning.
  • In the well-specified stochastic (monotone) setting, it proves an expected cumulative log-likelihood gap of order O(n^{1/3}) between the online estimators and the true density.
  • For the expert aggregation method, it also derives a pathwise regret bound of order √(n log n) relative to the best offline monotone estimator chosen in hindsight, with minimal assumptions on the data sequence.
  • As an application, it links log-optimal p-to-e calibration for sequential hypothesis testing to online monotone density estimation, and then builds empirically adaptive calibrators with theoretical optimality and supporting numerical experiments.

Abstract

We study the problem of online monotone density estimation, where density estimators must be constructed in a predictable manner from sequentially observed data. We propose two online estimators: an online analogue of the classical Grenander estimator, and an expert aggregation estimator inspired by exponential weighting methods from the online learning literature. In the well-specified stochastic setting, where the underlying density is monotone, we show that the expected cumulative log-likelihood gap between the online estimators and the true density admits an O(n^{1/3}) bound. We further establish a \sqrt{n\log{n}} pathwise regret bound for the expert aggregation estimator relative to the best offline monotone estimator chosen in hindsight, under minimal regularity assumptions on the observed sequence. As an application of independent interest, we show that the problem of constructing log-optimal p-to-e calibrators for sequential hypothesis testing can be formulated as an online monotone density estimation problem. We adapt the proposed estimators to build empirically adaptive p-to-e calibrators and establish their optimality. Numerical experiments illustrate the theoretical results.