AI Navigate

Historical Consensus: Preventing Posterior Collapse via Iterative Selection of Gaussian Mixture Priors

arXiv cs.LG / 3/12/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • It proposes Historical Consensus Training (HCT), a novel iterative procedure that selects and refines a set of Gaussian Mixture Model priors to prevent posterior collapse in variational autoencoders.
  • The approach leverages multiple clustering constraints to create a historical barrier in parameter space that remains stable under single-objective training.
  • The authors prove the historical barrier excludes collapsed solutions and demonstrate empirically that non-collapsed representations are achieved regardless of decoder variance or regularization strength.
  • The method requires no explicit stability conditions and is compatible with arbitrary neural architectures.
  • The code is publicly available at the provided GitHub repository: https://github.com/tsegoochang/historical-consensus-vae.

Abstract

Variational autoencoders (VAEs) frequently suffer from posterior collapse, where latent variables become uninformative and the approximate posterior degenerates to the prior. Recent work has characterized this phenomenon as a phase transition governed by the spectral properties of the data covariance matrix. In this paper, we propose a fundamentally different approach: instead of avoiding collapse through architectural constraints or hyperparameter tuning, we eliminate the possibility of collapse altogether by leveraging the multiplicity of Gaussian mixture model (GMM) clusterings. We introduce Historical Consensus Training, an iterative selection procedure that progressively refines a set of candidate GMM priors through alternating optimization and selection. The key insight is that models trained to satisfy multiple distinct clustering constraints develop a historical barrier -- a region in parameter space that remains stable even when subsequently trained with a single objective. We prove that this barrier excludes the collapsed solution, and demonstrate through extensive experiments on synthetic and real-world datasets that our method achieves non-collapsed representations regardless of decoder variance or regularization strength. Our approach requires no explicit stability conditions (e.g., \sigma^{\prime 2} < \lambda_{\max}) and works with arbitrary neural architectures. The code is available at https://github.com/tsegoochang/historical-consensus-vae.