Compression Favors Consistency, Not Truth: When and Why Language Models Prefer Correct Information
arXiv cs.CL / 3/13/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper introduces the Compression-Consistency Principle, arguing that next-token prediction favors hypotheses that yield shorter, internally consistent descriptions of the training data.
- It argues that truth bias in language models is not an intrinsic drive toward truth but arises when false alternatives are harder to compress structurally.
- In experiments with GPT-2–style models on synthetic data, correct completions reach 83.1% accuracy at balanced data and 67.0% when correct rules constitute only 10% of the corpus.
- Replacing random errors with a coherent but incorrect rule system largely eliminates the preference for correctness, driving accuracy toward chance, with a weaker but present effect in more natural-language-like synthetic settings (57.7%).
- The authors show that embedding verification steps can restore the preference for correctness at small scales and that more consistent rules yield graded accuracy improvements, suggesting that the observed truth bias is a byproduct of compression pressure rather than an intrinsic truth-seeking drive.
Related Articles

Attacks On Data Centers, Qwen3.5 In All Sizes, DeepSeek’s Huawei Play, Apple’s Multimodal Tokenizer
The Batch

Your AI generated code is "almost right", and that is actually WORSE than it being "wrong".
Dev.to

Lessons from Academic Plagiarism Tools for SaaS Product Development
Dev.to

**Core Allocation Optimization for Energy‑Efficient Multi‑Core Scheduling in ARINC650 Systems**
Dev.to

KI in der amtlichen Recherche beim DPMA: Was Patentanwälte bei Neuanmeldungen jetzt beachten sollten (Stand: März 2026)
Dev.to