A Linguistics-Aware LLM Watermarking via Syntactic Predictability

arXiv cs.CL / 4/20/2026

📰 NewsIdeas & Deep AnalysisTools & Practical UsageModels & Research

Key Points

  • The paper introduces STELA, a linguistics-aware watermarking framework for LLMs designed to improve the trade-off between text quality and watermark detectability.
  • STELA modulates watermark strength using POS n-gram–modeled linguistic indeterminacy, weakening the signal in grammatically constrained contexts and strengthening it where language is more flexible.
  • Unlike earlier approaches that require model-specific signals (such as access to logits), STELA’s detector can operate without any model logits, enabling publicly verifiable detection.
  • Experiments across typologically diverse languages (English, Chinese-only, and agglutinative Korean) indicate STELA outperforms prior watermarking methods in detection robustness.
  • The authors provide an implementation at the linked GitHub repository to support adoption and evaluation by others.

Abstract

As large language models (LLMs) continue to advance rapidly, reliable governance tools have become critical. Publicly verifiable watermarking is particularly essential for fostering a trustworthy AI ecosystem. A central challenge persists: balancing text quality against detection robustness. Recent studies have sought to navigate this trade-off by leveraging signals from model output distributions (e.g., token-level entropy); however, their reliance on these model-specific signals presents a significant barrier to public verification, as the detection process requires access to the logits of the underlying model. We introduce STELA, a novel framework that aligns watermark strength with the linguistic degrees of freedom inherent in language. STELA dynamically modulates the signal using part-of-speech (POS) n-gram-modeled linguistic indeterminacy, weakening it in grammatically constrained contexts to preserve quality and strengthening it in contexts with greater linguistic flexibility to enhance detectability. Our detector operates without access to any model logits, thus facilitating publicly verifiable detection. Through extensive experiments on typologically diverse languages-analytic English, isolating Chinese, and agglutinative Korean-we show that STELA surpasses prior methods in detection robustness. Our code is available at https://github.com/Shinwoo-Park/stela_watermark.