Causal Representation Learning from General Environments under Nonparametric Mixing

arXiv cs.LG / 4/28/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies causal representation learning, aiming to recover latent causal variables and their causal graph (a DAG) from raw observations using data from multiple environments.
  • It argues that many prior approaches rely on restrictive assumptions about how distributions mix or how causal mechanisms change, which often fail in real-world problems.
  • The authors introduce “general environments,” and show conditions under which the latent DAG and latent variables can be fully identified despite using a nonparametric mixing function and nonlinear causal models.
  • The key technical advance is leveraging “sufficient change conditions” on causal mechanisms, quantified using derivatives up to the third order, to achieve recovery up to only minor indeterminacies.
  • The work claims to be among the first results that fully recover the latent DAG under general environments with nonparametric mixing, while matching or improving existing results but with weaker assumptions.

Abstract

Causal representation learning aims to recover the latent causal variables and their causal relations, typically represented by directed acyclic graphs (DAGs), from low-level observations such as image pixels. A prevailing line of research exploits multiple environments, which assume how data distributions change, including single-node interventions, coupled interventions, or hard interventions, or parametric constraints on the mixing function or the latent causal model, such as linearity. Despite the novelty and elegance of the results, they are often violated in real problems. Accordingly, we formalize a set of desiderata for causal representation learning that applies to a broader class of environments, referred to as general environments. Interestingly, we show that one can fully recover the latent DAG and identify the latent variables up to minor indeterminacies under a nonparametric mixing function and nonlinear latent causal models, such as additive (Gaussian) noise models or heteroscedastic noise models, by properly leveraging sufficient change conditions on the causal mechanisms up to third-order derivatives. These represent, to our knowledge, the first results to fully recover the latent DAG from general environments under nonparametric mixing. Notably, our results match or improve upon many existing works, but require less restrictive assumptions about changing environments.