Exemplar Retrieval Without Overhypothesis Induction: Limits of Distributional Sequence Learning in Early Word Learning

arXiv cs.CL / 4/8/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The study investigates whether autoregressive transformer language models can learn the child-like “overhypothesis” that shape is a category-defining feature, using synthetic corpora that control for alternative explanations.
  • Across 120 pre-registered runs with 3.4M–25.6M parameter models, the systems reached perfect first-order exemplar retrieval on a large wug test battery but failed to achieve second-order generalization to novel nouns (performance stayed at chance).
  • An equivalence test and feature-swap diagnostic both support that the models’ behavior stems from frame-to-feature template matching rather than structured abstraction akin to noun→domain→feature reasoning.
  • The authors conclude that distributional sequence learning alone has a clear limitation for modeling developmental-scale mechanisms required for early word learning overgeneralizations.

Abstract

Background: Children do not simply learn that balls are round and blocks are square. They learn that shape is the kind of feature that tends to define object categories -- a second-order generalisation known as an overhypothesis [1, 2]. What kind of learning mechanism is sufficient for this inductive leap? Methods: We trained autoregressive transformer language models (3.4M-25.6M parameters) on synthetic corpora in which shape is the stable feature dimension across categories, with eight conditions controlling for alternative explanations. Results: Across 120 pre-registered runs evaluated on a 1,040-item wug test battery, every model achieved perfect first-order exemplar retrieval (100%) while second-order generalisation to novel nouns remained at chance (50-52%), a result confirmed by equivalence testing. A feature-swap diagnostic revealed that models rely on frame-to-feature template matching rather than structured noun-to-domain-to-feature abstraction. Conclusions: These results reveal a clear limitation of autoregressive distributional sequence learning under developmental-scale training conditions.