Diverging Transformer Predictions for Human Sentence Processing: A Comprehensive Analysis of Agreement Attraction Effects
arXiv cs.CL / 3/18/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The study systematically evaluates eleven autoregressive transformers using a surprisal-based linking mechanism across a broad set of English agreement attraction configurations to test their cognitive plausibility.
- The results show transformers generally align with human reading times for prepositional phrase configurations but fail to replicate human asymmetries in object-extracted relative clauses, with large model-to-model variation.
- No model fully reproduces the asymmetric interference patterns observed in humans, suggesting current transformers do not adequately explain human morphosyntactic processing.
- The authors argue that evaluating transformers as cognitive models requires rigorous, comprehensive experimental designs to avoid spurious generalizations from isolated syntactic configurations or single models.
Related Articles
Astral to Join OpenAI
Dev.to

PearlOS. We gave swarm intelligence a local desktop environment and code control to self-evolve. Has been pretty incredible to see so far. Open source and free if you want your own.
Reddit r/LocalLLaMA

The programming passion is melting
Dev.to

Maximize Developer Revenue with Monetzly's Innovative API for AI Conversations
Dev.to
Co-Activation Pattern Detection for Prompt Injection: A Mechanistic Interpretability Approach Using Sparse Autoencoders
Reddit r/LocalLLaMA