Phase transition on a context-sensitive random language model with short range interactions

arXiv stat.ML / 4/2/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies statistical-mechanics behavior of a random language model by explicitly constructing one with short-range (not long-range) symbol interactions.
  • It positions the model within context-sensitive grammars (Chomsky hierarchy), enabling explicit dependence on referenced contexts.
  • Through numerical investigation, the authors find a Berezinskii–Kosterlitz–Thouless–type phase transition persists even when the model only refers to contexts whose length stays constant as sentences grow.
  • The results suggest finite-temperature phase transitions in language models stem from intrinsic linguistic/context structure rather than being an artifact of long-range interactions like those in earlier long-range models.
  • The work extends the understanding of what mechanisms—beyond interaction range—can produce thermodynamic phase transitions in language-model-like systems.

Abstract

Since the random language model was proposed by E. DeGiuli [Phys. Rev. Lett. 122, 128301], language models have been investigated intensively from the viewpoint of statistical mechanics. Recently, the existence of a Berezinskii--Kosterlitz--Thouless transition was numerically demonstrated in models with long-range interactions between symbols. In statistical mechanics, it has long been known that long-range interactions can induce phase transitions. Therefore, it has remained unclear whether phase transitions observed in language models originate from genuinely linguistic properties that are absent in conventional spin models. In this study, we construct a random language model with short-range interactions and numerically investigate its statistical properties. Our model belongs to the class of context-sensitive grammars in the Chomsky hierarchy and allows explicit reference to contexts. We find that a phase transition occurs even when the model refers only to contexts whose length remains constant with respect to the sentence length. This result indicates that finite-temperature phase transitions in language models are genuinely induced by the intrinsic nature of language, rather than by long-range interactions.