Quantisation Reshapes the Metacognitive Geometry of Language Models
arXiv cs.CL / 4/13/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper finds that quantization changes LLM “metacognitive efficiency” by restructuring domain-level M-ratio behavior rather than uniformly degrading it.
- In experiments with Llama-3-8B-Instruct on 3,000 questions, M-ratio profiles across four knowledge domains are uncorrelated between Q5_K_M and f16 (Spearman rho = 0.00), with some domains improving while others worsen after quantization.
- Type-2 AUROC profiles remain perfectly stable across formats (rho = 1.00), suggesting the effect is mainly in M-ratio normalization/confidence calibration rather than the underlying discrimination signal.
- A pre-registered attempt to improve metacognition via domain-conditional confidence-amplification SFT did not generalize: confirmatory hypotheses were null, and meta-d’ did not improve because the diagnostic profile didn’t transfer across quantization formats.
- The authors release code, pre-registrations, and trial-level data and warn that systems relying on domain-level M-ratio profiles may have an unexamined dependency on inference format, while AUROC_2 may be safer.
Related Articles

Black Hat Asia
AI Business

I built the missing piece of the MCP ecosystem
Dev.to

When Agents Go Wrong: AI Accountability and the Payment Audit Trail
Dev.to

Google Gemma 4 Review 2026: The Open Model That Runs Locally and Beats Closed APIs
Dev.to

OpenClaw Deep Dive Guide: Self-Host Your Own AI Agent on Any VPS (2026)
Dev.to