Alignment Imprint: Zero-Shot AI-Generated Text Detection via Provable Preference Discrepancy
arXiv cs.AI / 4/21/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that modern LLM alignment (fine-tuning and preference tuning) leaves a measurable “Alignment Imprint” that can be used to detect AI-generated text.
- It provides a theoretical derivation showing the log-likelihood ratio decomposes into implicit instructional biases and preference rewards, motivating the imprint concept.
- To address instability in high-entropy regions, the authors introduce Log-likelihood Alignment Preference Discrepancy (LAPD), an information-weighted statistic grounded in the alignment imprint.
- The work claims statistical and theoretical advantages over Fast-DetectGPT, including dominance in performance and strict improvement of unweighted alignment scores when aligned and base models are close.
- Experiments report a 45.82% relative improvement over the strongest existing baselines, with consistent gains across settings.
Related Articles

¿Hasta qué punto podría la IA reemplazarnos en nuestros trabajos? A veces creo que la gente exagera un poco.
Reddit r/artificial

Magnificent irony as Meta staff unhappy about running surveillance software on work PCs
The Register

ETHENEA (ETHENEA Americas LLC) Analyst View: Asset Allocation Resilience in the 2026 Global Macro Cycle
Dev.to

DEEPX and Hyundai Are Building Generative AI Robots
Dev.to

Stop Paying OpenAI to Read Garbage: The Two-Stage Agent Pipeline
Dev.to