Forecasting Multivariate Time Series under Predictive Heterogeneity: A Validation-Driven Clustering Framework

arXiv stat.ML / 4/16/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses multivariate time series forecasting when predictive heterogeneity makes global pooling statistically efficient but potentially mismatched to different underlying predictive regimes.
  • It proposes an adaptive pooling approach cast as a statistical decision problem that uses out-of-sample validation performance to decide when and how specialization (clustering) should be applied.
  • Instead of clustering by representation similarity, partitions are defined via predictive risk (expected out-of-sample loss) and approximated with validation error.
  • Cluster assignments are updated iteratively using validation losses for both point forecasting (Huber loss) and probabilistic forecasting (pinball loss) to improve robustness to heavy-tailed errors and local anomalies.
  • A leakage-free fallback mechanism ensures the system reverts to a global model when specialization does not improve validation performance, preventing degradation under a strict train/validation/test protocol.

Abstract

We study adaptive pooling under predictive heterogeneity in high-dimensional multivariate time series forecasting, where global models improve statistical efficiency but may fail to capture heterogeneous predictive structure, while naive specialization can induce negative transfer. We formulate adaptive pooling as a statistical decision problem and propose a validation-driven framework that determines when and how specialization should be applied. Rather than grouping series based on representation similarity, we define partitions through out-of-sample predictive performance, thereby aligning data organization with predictive risk, defined as expected out-of-sample loss and approximated via validation error. Cluster assignments are iteratively updated using validation losses for both point (Huber) and probabilistic (pinball) forecasting, improving robustness to heavy-tailed errors and local anomalies. To ensure reliability, we introduce a leakage-free fallback mechanism that reverts to a global model whenever specialization fails to improve validation performance, providing a safeguard against performance degradation under a strict training-validation-test protocol. Experiments on large-scale traffic datasets demonstrate consistent improvements over strong baselines while avoiding degradation when heterogeneity is weak. Overall, the proposed framework provides a principled and practically reliable approach to adaptive pooling in high-dimensional forecasting problems.