PDGMM-VAE: A Variational Autoencoder with Adaptive Per-Dimension Gaussian Mixture Model Priors for Nonlinear ICA

arXiv stat.ML / 3/26/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper introduces PDGMM-VAE, a source-oriented variational autoencoder for blind source separation and nonlinear ICA, where each latent dimension is treated as an individual source signal.
  • Instead of using a single shared prior, the method assigns an adaptive Gaussian mixture model (GMM) prior per latent dimension to better match heterogeneous non-Gaussian source statistics.
  • The GMM prior parameters are not predefined; they are learned jointly with the encoder and decoder via end-to-end training, with refinement toward convergence under the overall objective.
  • In the proposed probabilistic encoder-decoder framework, the encoder functions as a demixing map from observations to inferred sources, while the decoder reconstructs the observed mixtures.
  • Experiments on both linear and nonlinear mixing scenarios report that PDGMM-VAE can recover latent sources and produce satisfactory separation performance.

Abstract

Independent component analysis is a core framework within blind source separation for recovering latent source signals from observed mixtures under statistical independence assumptions. In this work, we propose PDGMM-VAE, a source-oriented variational autoencoder in which each latent dimension, interpreted explicitly as an individual source signal, is assigned its own Gaussian mixture model prior. Unlike conventional VAE formulations with a shared simple prior, the proposed framework imposes per-dimension heterogeneous prior constraints, enabling the model to capture diverse non-Gaussian source statistics and thereby promote source separation under a probabilistic encoder-decoder architecture. Importantly, the parameters of these per-dimension GMM priors are not fixed in advance, but are adaptively learned and automatically refined toward convergence together with the encoder and decoder parameters under the overall training objective. Within this formulation, the encoder serves as a demixing mapping from observations to latent sources, while the decoder reconstructs the observed mixtures from the inferred components. The proposed model provides a systematic study of an idea that had previously only been noted in our preliminary form, namely, equipping different latent sources with different GMM priors for ICA, and formulates it as a full VAE framework with end-to-end training and per-dimension prior learning. Experimental results on both linear and nonlinear mixing problems demonstrate that PDGMM-VAE can recover latent source signals and achieve satisfactory separation performance.