Embracing Heteroscedasticity for Probabilistic Time Series Forecasting

arXiv cs.LG / 3/26/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • Probabilistic time series forecasting (PTSF) requires modeling not just point predictions but the full predictive distribution, making time-varying (heteroscedastic) uncertainty a core need.
  • The paper argues that many non-autoregressive generative PTSF methods trained with MSE implicitly assume homoscedastic noise, limiting their ability to learn temporal heteroscedasticity.
  • It introduces Location-Scale Gaussian VAE (LSG-VAE), which explicitly parameterizes both predictive mean (location) and time-dependent variance (scale) via a location-scale likelihood.
  • LSG-VAE is reported to capture heteroscedastic aleatoric uncertainty and uses an adaptive attenuation mechanism to down-weight highly volatile observations during training for better robustness.
  • Experiments across nine benchmark datasets show LSG-VAE outperforming fifteen generative baselines while remaining computationally efficient for real-time deployment.

Abstract

Probabilistic time series forecasting (PTSF) aims to model the full predictive distribution of future observations, enabling both accurate forecasting and principled uncertainty quantification. A central requirement of PTSF is to embrace heteroscedasticity, as real-world time series exhibit time-varying conditional variances induced by nonstationary dynamics, regime changes, and evolving external conditions. However, most existing non-autoregressive generative approaches to PTSF, such as TimeVAE and K^2VAE, rely on MSE-based training objectives that implicitly impose a homoscedastic assumption, thereby fundamentally limiting their ability to model temporal heteroscedasticity. To address this limitation, we propose the Location-Scale Gaussian VAE (LSG-VAE), a simple but effective framework that explicitly parameterizes both the predictive mean and time-dependent variance through a location-scale likelihood formulation. This design enables LSG-VAE to faithfully capture heteroscedastic aleatoric uncertainty and introduces an adaptive attenuation mechanism that automatically down-weights highly volatile observations during training, leading to improved robustness in trend prediction. Extensive experiments on nine benchmark datasets demonstrate that LSG-VAE consistently outperforms fifteen strong generative baselines while maintaining high computational efficiency suitable for real-time deployment.