A Linguistics-Aware LLM Watermarking via Syntactic Predictability
arXiv cs.CL / 4/20/2026
📰 NewsIdeas & Deep AnalysisTools & Practical UsageModels & Research
Key Points
- The paper introduces STELA, a linguistics-aware watermarking framework for LLMs designed to improve the trade-off between text quality and watermark detectability.
- STELA modulates watermark strength using POS n-gram–modeled linguistic indeterminacy, weakening the signal in grammatically constrained contexts and strengthening it where language is more flexible.
- Unlike earlier approaches that require model-specific signals (such as access to logits), STELA’s detector can operate without any model logits, enabling publicly verifiable detection.
- Experiments across typologically diverse languages (English, Chinese-only, and agglutinative Korean) indicate STELA outperforms prior watermarking methods in detection robustness.
- The authors provide an implementation at the linked GitHub repository to support adoption and evaluation by others.
💡 Insights using this article
This article is featured in our daily AI news digest — key takeaways and action items at a glance.



