Gaussian mixture models as a proxy for interacting language models

arXiv stat.ML / 4/7/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes interacting Gaussian mixture models (GMMs) as a computationally cheap proxy for studying interactions between large language models (LLMs).
  • It defines an interacting GMM system that includes an analogue to retrieval-augmented generation (RAG) via an updating mechanism for data and parameters exchanged among components.
  • The authors show that this GMM interaction framework can mimic certain behaviors observed in simulations of interacting LLMs that iteratively respond with feedback from other models.
  • They construct a Markov chain representation of the interacting GMMs, formalize polarization within that chain, and prove lower bounds on the probability of polarization.
  • Overall, the work provides theoretical insight into when and how interacting GMMs can approximate qualitative dynamics of interacting LLM systems without the heavy computational cost.

Abstract

Large language models (LLMs) are powerful tools that, in a number of settings, overlap with the results of human pattern recognition and reasoning. Retrieval-augmented generation (RAG) further allows LLMs to produce tailored output depending on the contents of their RAG databases. However, LLMs depend on complex, computationally expensive algorithms. In this paper, we introduce interacting Gaussian mixture models (GMMs) as a proxy for interacting LLMs. We construct a model of interacting GMMs, complete with an analogue to RAG updating, under which GMMs can generate, exchange, and update data and parameters. We show that this interacting system of Gaussian mixture models, which can be implemented at minimal computational cost, mimics certain aspects of experimental simulations of interacting LLMs whose iterative responses depend on feedback from other LLMs. We build a Markov chain from this system of interacting GMMs; formalize and interpret the notion of polarization for such a chain; and prove lower bounds on the probability of polarization. This provides theoretical insight into the use of interacting Gaussian mixture models as a computationally efficient proxy for interacting large language models.