LLM Psychosis: A Theoretical and Diagnostic Framework for Reality-Boundary Failures in Large Language Models
arXiv cs.AI / 4/30/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that “hallucination” is an inadequate label for a distinct class of LLM failures seen when models act as interactive agents, proposing “LLM Psychosis” as a structured framework.
- It defines five hallmark features—reality-boundary dissolution, persistence of injected false beliefs, logical incoherence under impossible constraints, self-model instability, and epistemic overconfidence—that it claims are qualitatively different from ordinary factual errors.
- To operationalize the framework, the authors introduce the LLM Cognitive Integrity Scale (LCIS), a five-axis diagnostic instrument covering ERI, PAI, LCR, SMI, and ECI.
- Using an adversarial probe battery on ChatGPT 5 (GPT-5), they report measurable integrity baselines and specific “psychosis-like” failure signatures across the LCIS axes.
- The results yield a three-tier severity taxonomy (Type I Confabulatory, Type II Delusional, Type III Dissociative) and identify a “delusional gradient” dynamic—where correction pressure can intensify the failure—as particularly consequential for safety evaluation and high-stakes screening.
Related Articles
Claude Opus 4.7: What Actually Changed and Whether You Should Migrate
Dev.to
Big Tech firms are accelerating AI investments and integration, while regulators and companies focus on safety and responsible adoption.
Dev.to
The Inference Inflection: Why AI's Center of Gravity Has Shifted from Training to Inference
Dev.to
AI transparency index on pvgomes.com
Dev.to
Mastering On-Device GenAI: How to Fine-Tune LLMs for Android Using LoRA and Kotlin 2.x
Dev.to