Language Diffusion Models are Associative Memories Capable of Retrieving Unseen Data

arXiv cs.LG / 4/30/2026

📰 NewsModels & Research

Key Points

  • The paper argues that Uniform-based Discrete Diffusion Models (UDDMs) can be understood as associative memories, with stable “attraction basins” emerging to recover stored data.
  • It proposes that explicit energy functions are not strictly required, because similar basin formation can arise from conditional likelihood maximization.
  • By comparing token recovery for training versus test examples, the authors find a sharp memorization-to-generalization transition controlled by training dataset size.
  • They show that this transition can be detected using only conditional entropy of predicted token sequences, where memorization corresponds to near-zero conditional entropy.
  • The results suggest a practical diagnostic for deployed language diffusion models to assess whether they are memorizing or truly generalizing.

Abstract

When do language diffusion models memorize their training data, and how to quantitatively assess their true generative regime? We address these questions by showing that Uniform-based Discrete Diffusion Models (UDDMs) fundamentally behave as Associative Memories (AMs) \textit{with emergent creative capabilities}. The core idea of an AM is to reliably recover stored data points as \textit{memories} by establishing distinct basins of attraction around them. Historically, models like Hopfield networks use an explicit energy function to guarantee these stable attractors. We broaden this perspective by leveraging the observation that energy is not strictly necessary, as basins of attraction can also be formed via conditional likelihood maximization. By evaluating token recovery of \textit{training} and \textit{test} examples, we identify in UDDMs a sharp memorization-to-generalization transition governed by the size of the training dataset: as it increases, basins around training examples shrink and basins around unseen test examples expand, until both later converge to the same level. Crucially, we can detect this transition using only the conditional entropy of predicted token sequences: memorization is characterized by vanishing conditional entropy, while in the generalization regime the conditional entropy of most tokens remains finite. Thus, conditional entropy offers a practical probe for the memorization-to-generalization transition in deployed models.