AI Navigate

Diverging Transformer Predictions for Human Sentence Processing: A Comprehensive Analysis of Agreement Attraction Effects

arXiv cs.CL / 3/18/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The study systematically evaluates eleven autoregressive transformers using a surprisal-based linking mechanism across a broad set of English agreement attraction configurations to test their cognitive plausibility.
  • The results show transformers generally align with human reading times for prepositional phrase configurations but fail to replicate human asymmetries in object-extracted relative clauses, with large model-to-model variation.
  • No model fully reproduces the asymmetric interference patterns observed in humans, suggesting current transformers do not adequately explain human morphosyntactic processing.
  • The authors argue that evaluating transformers as cognitive models requires rigorous, comprehensive experimental designs to avoid spurious generalizations from isolated syntactic configurations or single models.

Abstract

Transformers underlie almost all state-of-the-art language models in computational linguistics, yet their cognitive adequacy as models of human sentence processing remains disputed. In this work, we use a surprisal-based linking mechanism to systematically evaluate eleven autoregressive transformers of varying sizes and architectures on a more comprehensive set of English agreement attraction configurations than prior work. Our experiments yield mixed results: While transformer predictions generally align with human reading time data for prepositional phrase configurations, performance degrades significantly on object-extracted relative clause configurations. In the latter case, predictions also diverge markedly across models, and no model successfully replicates the asymmetric interference patterns observed in humans. We conclude that current transformer models do not explain human morphosyntactic processing, and that evaluations of transformers as cognitive models must adopt rigorous, comprehensive experimental designs to avoid spurious generalizations from isolated syntactic configurations or individual models.