Inference of Multiscale Gaussian Graphical Model

arXiv stat.ML / 3/25/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes a Multiscale Graphical Lasso (MGLasso) that performs clustering and conditional-independence graph inference simultaneously for Gaussian Graphical Models (GGMs).
  • MGLasso improves interpretability by learning graphs at multiple granularity levels using a convex clustering approach (relaxations of k-means and hierarchical clustering) together with neighborhood selection for undirected graphs.
  • It extends and generalizes sparse group fused lasso formulations to undirected graphical models, enabling joint estimation of cluster structure and network edges.
  • The authors develop an optimization approach using CONESTA (continuation with Nesterov smoothing plus shrinkage-thresholding) to compute a regularization path efficiently along a group fused-lasso penalty while holding the Lasso penalty constant.
  • Experiments on synthetic data and real applications (gut microbiome and poplar methylation with transcriptomics) show comparative performance against existing clustering and network inference methods.

Abstract

Gaussian Graphical Models (GGMs) are widely used in high-dimensional data analysis to synthesize the interaction between variables. In many applications, such as genomics or image analysis, graphical models rely on sparsity and clustering to reduce dimensionality and improve performances. This paper explores a slightly different paradigm where clustering is not knowledge-driven but performed simultaneously with the graph inference task. We introduce a novel Multiscale Graphical Lasso (MGLasso) to improve networks interpretability by proposing graphs at different granularity levels. The method estimates clusters through a convex clustering approach - a relaxation of k-means, and hierarchical clustering. The conditional independence graph is simultaneously inferred through a neighborhood selection scheme for undirected graphical models. MGLasso extends and generalizes the sparse group fused lasso problem to undirected graphical models. We use continuation with Nesterov smoothing in a shrinkage-thresholding algorithm (CONESTA) to propose a regularization path of solutions along the group fused Lasso penalty, while the Lasso penalty is kept constant. Extensive experiments on synthetic data compare the performances of our model to state-of-the-art clustering methods and network inference models. Applications to gut microbiome data and poplar's methylation mixed with transcriptomic data are presented.