A Foundation Model for Instruction-Conditioned In-Context Time Series Tasks

arXiv cs.LG / 3/25/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a time-series foundation model that supports instruction-conditioned in-context learning, using demonstrations rather than task-specific fine-tuning.
  • It builds an encoder-decoder (quantile-regression T5) with structured tokenization that explicitly marks target series, covariates, context, and task-specific future information.
  • A hierarchical Transformer architecture performs per-example encoding and cross-example attention during decoding to condition forecasts on demonstration pairs.
  • The model is trained on large-scale real and synthetic data with supervised forecasting plus multiple self-supervised tasks (imputation, reconstruction, classification, anomaly detection, and source demixing) to learn mappings across tasks.
  • Experiments across datasets, frequencies, and horizons show improved performance over strong time-series foundation baselines on point and probabilistic forecasting benchmarks, while staying competitive for classification and anomaly detection.

Abstract

In-context learning (ICL) allows a model to adapt at inference time by conditioning on examples rather than updating parameters. Existing time-series foundation models use implicit positional context, retrieval, or task-specific objectives, but rarely explicit instruction-conditioned demonstrations. We present a foundation model for instruction-conditioned in-context time-series tasks based on a quantile-regression T5 encoder-decoder. Historical examples and queries are encoded with a structured tokenization scheme that marks target series, covariates, context, and task-specific future information. A hierarchical Transformer with per-example encoding, example-level fusion, and cross-example attention conditions decoding on demonstration pairs, enabling forecasting and related tasks without task-specific fine-tuning. We train on large-scale real and synthetic time series using supervised forecasting plus self-supervised tasks, including imputation, reconstruction, classification, anomaly detection, and source demixing. This multi-task training learns a distribution over task mappings and improves adaptation to local structure at inference time. Across diverse datasets, frequencies, and horizons, our method outperforms strong foundation baselines on point and probabilistic forecasting benchmarks, including fev-bench and GIFT-Eval, while remaining competitive on classification and anomaly detection.