Is This Just Fantasy? Language Model Representations Reflect Human Judgments of Event Plausibility
arXiv cs.CL / 4/29/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper investigates whether language models can accurately categorize sentence modality (e.g., possible, impossible, nonsensical) as required by many downstream tasks.
- It identifies “modal difference vectors” (linear representations) within multiple LMs that distinguish between modal categories more reliably than prior studies suggested.
- The authors show that these modal difference vectors appear in a consistent progression as models improve with training, depth (layers), and parameter scaling.
- They demonstrate that directions in activation space can predict fine-grained human judgments of event plausibility, linking model internal representations to interpretable features used by people.
- The work uses mechanistic interpretability techniques to provide new insights that may help explain how humans process and distinguish modal categories.
Related Articles

How I Use AI Agents to Maintain a Living Knowledge Base for My Team
Dev.to
IK_LLAMA now supports Qwen3.5 MTP Support :O
Reddit r/LocalLLaMA
OpenAI models, Codex, and Managed Agents come to AWS
Dev.to

Indian Developers: How to Build AI Side Income with $0 Capital in 2026
Dev.to

Vertical SaaS for Startups 2026: Building a Niche AI-First Product
Dev.to