How Transformers Reject Wrong Answers: Rotational Dynamics of Factual Constraint Processing
arXiv cs.CL / 3/17/2026
📰 NewsIdeas & Deep AnalysisModels & Research
Key Points
- It introduces forced-completion probing to compare identical queries with correct and incorrect single-token continuations across all layers of four decoder-only models (1.5B-13B parameters).
- It shows that correct and incorrect paths diverge via rotation on an approximate hypersphere, with displacement magnitudes staying similar while angular separation grows across layers.
- It finds that models actively suppress the correct answer when faced with incorrect input, moving probability away from the right token rather than passively failing.
- It observes a parameter threshold around 1.6B where these effects emerge, indicating a phase-transition in factual processing capability.
Related Articles

Astral to Join OpenAI
Dev.to

PearlOS. We gave swarm intelligence a local desktop environment and code control to self-evolve. Has been pretty incredible to see so far. Open source and free if you want your own.
Reddit r/LocalLLaMA

Why Data is Important for LLM
Dev.to

The Inference Market Is Consolidating. Agent Payments Are Still Nobody's Problem.
Dev.to

YouTube's Deepfake Shield for Politicians Changes Evidence Forever
Dev.to