AI Navigate

Interventional Time Series Priors for Causal Foundation Models

arXiv cs.LG / 3/13/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • It presents CausalTimePrior, a principled framework for generating synthetic temporal structural causal models with both observational and interventional time series, addressing the lack of interventional data in benchmarks.
  • The framework supports configurable causal graph structures, nonlinear autoregressive mechanisms, regime-switching dynamics, and multiple intervention types including hard, soft, and time-varying interventions.
  • The study shows PFNs trained on data generated by CausalTimePrior can perform in-context causal effect estimation on held-out TSCMs, signaling progress toward time-series causal foundation models.
  • By providing synthetic interventional targets, the approach creates a path toward training and evaluating foundation models for time-series causal inference, bridging a gap in current benchmarks.

Abstract

Prior-data fitted networks (PFNs) have emerged as powerful foundation models for tabular causal inference, yet their extension to time series remains limited by the absence of synthetic data generators that provide interventional targets. Existing time series benchmarks generate observational data with ground-truth causal graphs but lack the interventional data required for training causal foundation models. To address this, we propose \textbf{CausalTimePrior}, a principled framework for generating synthetic temporal structural causal models (TSCMs) with paired observational and interventional time series. Our prior supports configurable causal graph structures, nonlinear autoregressive mechanisms, regime-switching dynamics, and multiple intervention types (hard, soft, time-varying). We demonstrate that PFNs trained on CausalTimePrior can perform in-context causal effect estimation on held-out TSCMs, establishing a pathway toward foundation models for time series causal inference.