Differentially Private and Federated Structure Learning in Bayesian Networks

arXiv stat.ML / 4/13/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes Fed-Sparse-BNSL, a federated algorithm for learning Bayesian network structure from decentralized linear Gaussian data while providing formal differential privacy guarantees.
  • It addresses the typical tradeoff between privacy and communication efficiency by performing greedy updates that only target a small number of relevant edges per participant, reducing communication costs as dimensionality grows.
  • The method is designed to preserve Bayesian network identifiability, enabling accurate structure estimation even under privacy constraints.
  • Experiments on both synthetic and real datasets show utility close to non-private baselines, alongside substantially stronger privacy protection and better communication efficiency.
  • Overall, the work advances privacy-preserving, communication-efficient structure learning for Bayesian networks in federated settings, which can enable safer analytics across multiple parties.

Abstract

Learning the structure of a Bayesian network from decentralized data poses two major challenges: (i) ensuring rigorous privacy guarantees for participants, and (ii) avoiding communication costs that scale poorly with dimensionality. In this work, we introduce Fed-Sparse-BNSL, a novel federated method for learning linear Gaussian Bayesian network structures that addresses both challenges. By combining differential privacy with greedy updates that target only a few relevant edges per participant, Fed-Sparse-BNSL efficiently uses the privacy budget while keeping communication costs low. Our careful algorithmic design preserves model identifiability and enables accurate structure estimation. Experiments on synthetic and real datasets demonstrate that Fed-Sparse-BNSL achieves utility close to non-private baselines while offering substantially stronger privacy and communication efficiency.