A Knowledge-Informed Pretrained Model for Causal Discovery

arXiv cs.LG / 2026/3/24

📰 ニュースIdeas & Deep AnalysisModels & Research

要点

  • The paper proposes a knowledge-informed pretrained model for causal discovery designed to work when only coarse domain knowledge is available, avoiding the need for costly interventions or strong ground-truth priors.
  • It uses a dual source encoder-decoder architecture to incorporate weak prior knowledge into causal discovery from observational data in a principled way.
  • The authors introduce a diverse pretraining dataset and a curriculum learning strategy to adapt the model across different prior strengths, graph densities, and variable scales.
  • Experiments across in-distribution, out-of-distribution, and real-world datasets show consistent improvements over baseline methods, along with strong robustness and practical applicability.

Abstract

Causal discovery has been widely studied, yet many existing methods rely on strong assumptions or fall into two extremes: either depending on costly interventional signals or partial ground truth as strong priors, or adopting purely data driven paradigms with limited guidance, which hinders practical deployment. Motivated by real-world scenarios where only coarse domain knowledge is available, we propose a knowledge-informed pretrained model for causal discovery that integrates weak prior knowledge as a principled middle ground. Our model adopts a dual source encoder-decoder architecture to process observational data in a knowledge-informed way. We design a diverse pretraining dataset and a curriculum learning strategy that smoothly adapts the model to varying prior strengths across mechanisms, graph densities, and variable scales. Extensive experiments on in-distribution, out-of distribution, and real-world datasets demonstrate consistent improvements over existing baselines, with strong robustness and practical applicability.