Noise Titration: Exact Distributional Benchmarking for Probabilistic Time Series Forecasting
arXiv stat.ML / 2026/3/24
💬 オピニオンIdeas & Deep AnalysisModels & Research
要点
- The paper argues that conventional time-series forecasting evaluation is largely unfalsifiable because it relies on passive observation of single historical trajectories rather than controllable perturbations.
- It proposes “Noise Titration,” an interventionist benchmarking method that injects calibrated Gaussian observation noise into systems with known dynamics to enable exact distributional evaluation via negative log-likelihood and calibrated distributional tests.
- The work extends the Fern architecture into a probabilistic generative model that directly parameterizes covariance structures on the Symmetric Positive Definite (SPD) cone to improve calibrated joint forecasting without expensive Jacobian modeling.
- Experiments suggest that state-of-the-art zero-shot foundation models fail under non-stationary regime shifts and higher noise due to context-parroting, while Fern better preserves the invariant measure and multivariate geometry for sharper calibration.

