Nonlinear Causal Discovery through a Sequential Edge Orientation Approach

arXiv stat.ML / 4/24/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses limitations of existing nonlinear causal discovery methods by proposing a sequential approach to orient edges in a CPDAG using pairwise additive noise models (PANM).
  • It proves that the sequential orientation procedure can recover the true causal DAG under a restricted additive noise model (ANM) assumption.
  • The authors introduce a new constraint-based algorithm that ranks undirected edges by how well they satisfy the PANM to define an evaluation order for edge directions.
  • For each candidate edge, they use a statistical test comparing log-likelihoods of competing directions on a subgraph induced by the candidate nodes and their identified parents.
  • Experiments on synthetic and real-world data show the method is computationally efficient, robust to model misspecification, and outperforms many existing nonlinear DAG learning approaches.

Abstract

Recent advances have established the identifiability of a directed acyclic graph (DAG) under additive noise models (ANMs), spurring the development of various causal discovery methods. However, most existing methods make restrictive model assumptions, rely heavily on general independence tests, or require substantial computational time. To address these limitations, we propose a sequential procedure to orient undirected edges in a completed partial DAG (CPDAG), representing an equivalence class of DAGs, by leveraging the pairwise additive noise model (PANM) to identify their causal directions. We prove that this procedure can recover the true causal DAG assuming a restricted ANM. Building on this result, we develop a novel constraint-based algorithm for learning causal DAGs under nonlinear ANMs. Given an estimated CPDAG, we develop a ranking procedure that sorts undirected edges by their adherence to the PANM, which defines an evaluation order of the edges. To determine the edge direction, we devise a statistical test that compares the log-likelihood values, evaluated with respect to the competing directions, of a sub-graph comprising just the candidate nodes and their identified parents in the partial DAG. We further establish the structural learning consistency of our algorithm in the large-sample limit. Extensive experiments on synthetic and real-world datasets demonstrate that our method is computationally efficient, robust to model misspecification, and consistently outperforms many existing nonlinear DAG learning methods.