How Transformers Reject Wrong Answers: Rotational Dynamics of Factual Constraint Processing
arXiv cs.CL / 3/17/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- It introduces forced-completion probing to compare identical queries with correct and incorrect single-token continuations across all layers of four decoder-only models (1.5B-13B parameters).
- It shows that correct and incorrect paths diverge via rotation on an approximate hypersphere, with displacement magnitudes staying similar while angular separation grows across layers.
- It finds that models actively suppress the correct answer when faced with incorrect input, moving probability away from the right token rather than passively failing.
- It observes a parameter threshold around 1.6B where these effects emerge, indicating a phase-transition in factual processing capability.
Related Articles
The massive shift toward edge computing and local processing
Dev.to
Self-Refining Agents in Spec-Driven Development
Dev.to
Week 3: Why I'm Learning 'Boring' ML Before Building with LLMs
Dev.to
The Three-Agent Protocol Is Transferable. The Discipline Isn't.
Dev.to

has anyone tried this? Flash-MoE: Running a 397B Parameter Model on a Laptop
Reddit r/LocalLLaMA