CausalCompass: Evaluating the Robustness of Time-Series Causal Discovery in Misspecified Scenarios

arXiv stat.ML / 5/1/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces CausalCompass, a benchmark framework to evaluate time-series causal discovery (TSCD) methods when modeling assumptions are violated, addressing the lack of robustness-focused evaluation in existing benchmarks.
  • Extensive experiments across eight assumption-violation scenarios show that no single TSCD method performs best in all settings.
  • Across varied scenarios, the strongest overall performers are “almost invariably” deep learning–based approaches, supported by hyperparameter sensitivity analyses and ablation studies.
  • An additional finding is that NTS-NOTEARS depends heavily on standardized preprocessing: it performs poorly in the default (vanilla) setting but improves substantially after standardization.
  • The authors provide an implementation, documentation, and datasets to help researchers and practitioners systematically test TSCD robustness for broader real-world adoption.

Abstract

Causal discovery from time series is a fundamental task in machine learning. However, its widespread adoption is hindered by a reliance on untestable causal assumptions and by the lack of robustness-oriented evaluation in existing benchmarks. To address these challenges, we propose CausalCompass, a flexible and extensible benchmark framework designed to assess the robustness of time-series causal discovery (TSCD) methods under violations of modeling assumptions. To demonstrate the practical utility of CausalCompass, we conduct extensive benchmarking of representative TSCD algorithms across eight assumption-violation scenarios. Our experimental results indicate that no single method consistently attains optimal performance across all settings. Nevertheless, the methods exhibiting superior overall performance across diverse scenarios are almost invariably deep learning-based approaches. We further provide hyperparameter sensitivity analyses to deepen the understanding of these findings. We additionally conduct ablation experiments to explain the strong performance of deep learning-based methods under assumption violations. We also find, somewhat surprisingly, that NTS-NOTEARS relies heavily on standardized preprocessing in practice, performing poorly in the vanilla setting but exhibiting strong performance after standardization. Finally, our work aims to provide a comprehensive and systematic evaluation of TSCD methods under assumption violations, thereby facilitating their broader adoption in real-world applications. The user-friendly implementation, documentation and datasets are available at https://anonymous.4open.science/r/CausalCompass-anonymous-5B4F/.