CombinationTS: A Modular Framework for Understanding Time-Series Forecasting Models

arXiv cs.LG / 5/5/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The article argues that time-series forecasting research should move from simply selecting models to performing modular attribution to determine which components genuinely drive performance.
  • It introduces CombinationTS, a probabilistic, self-contained evaluation framework that decomposes forecasting architectures into orthogonal modules (Input Transformation, Embedding, Encoder, Decoder, Output Transformation) and tests them under shared evaluation conditions.
  • CombinationTS quantifies each module’s contribution using marginalized performance (μ) and stability (σ), aiming to produce robust, less fragile conclusions than point estimates.
  • Using large-scale paired evaluations, the authors report the “Identity Paradox,” where a parameter-free Identity Encoder can match or beat complex encoders when embeddings are well designed.
  • The study also finds that adding explicit structural priors via Input Transformations can improve the performance–stability trade-off more effectively than increasing Encoder complexity, motivating a principled baseline for architecture design.

Abstract

Recent progress in time-series forecasting has led to rapidly increasing architectural complexity, yet many reported State-of-the-Art gains are statistically fragile or misattributed. We argue that progress requires a shift from model selection to modular attribution, identifying which components truly drive performance. We propose CombinationTS, a self-contained probabilistic evaluation framework that decomposes forecasting models into orthogonal modules--Input Transformation, Embedding, Encoder, Decoder, and Output Transformation--and evaluates them under a shared evaluation condition space. By quantifying each component via marginalized performance (\mu) and stability (\sigma), CombinationTS enables robust attribution beyond fragile point estimates. Through large-scale paired evaluation, we uncover the Identity Paradox: once the data view (Embedding) is well-designed, a parameter-free Identity Encoder often matches or outperforms complex backbones. We further show that explicit structural priors introduced via Input Transformations yield a more favorable performance-stability trade-off than increasing Encoder complexity, establishing a principled baseline for architectural necessity.