Differentially Private Language Generation and Identification in the Limit

arXiv stat.ML / 4/10/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper studies Kleinberg and Mullainathan’s “language generation in the limit” model under differential privacy constraints using a continual release setting.
  • It shows that for any countable collection of languages, there is an ε-differentially private in-limit generator with no qualitative loss in feasibility, though privacy increases the required number of samples.
  • It proves strong impossibility results for ε-DP language identification in the limit when the target language collection has two languages with an infinite intersection and finite set difference.
  • In a stochastic (i.i.d. sampling) variant, the authors characterize when private identification is possible, showing it aligns with adversarial identifiability but diverges under privacy constraints.
  • Overall, the work delineates new theoretical differences between generation and identification under differential privacy, including separations between adversarial and stochastic regimes for identification.

Abstract

We initiate the study of language generation in the limit, a model recently introduced by Kleinberg and Mullainathan [KM24], under the constraint of differential privacy. We consider the continual release model, where a generator must eventually output a stream of valid strings while protecting the privacy of the entire input sequence. Our first main result is that for countable collections of languages, privacy comes at no qualitative cost: we provide an \varepsilon-differentially-private algorithm that generates in the limit from any countable collection. This stands in contrast to many learning settings where privacy renders learnability impossible. However, privacy does impose a quantitative cost: there are finite collections of size k for which uniform private generation requires \Omega(k/\varepsilon) samples, whereas just one sample suffices non-privately. We then turn to the harder problem of language identification in the limit. Here, we show that privacy creates fundamental barriers. We prove that no \varepsilon-DP algorithm can identify a collection containing two languages with an infinite intersection and a finite set difference, a condition far stronger than the classical non-private characterization of identification. Next, we turn to the stochastic setting where the sample strings are sampled i.i.d. from a distribution (instead of being generated by an adversary). Here, we show that private identification is possible if and only if the collection is identifiable in the adversarial model. Together, our results establish new dimensions along which generation and identification differ and, for identification, a separation between adversarial and stochastic settings induced by privacy constraints.