Transformer See, Transformer Do: Copying as an Intermediate Step in Learning Analogical Reasoning
arXiv cs.LG / 4/9/2026
💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- The paper investigates how to train transformer models to perform analogical reasoning on letter-string analogy tasks using meta-learning for compositionality (MLC).
- It finds that adding explicit copying tasks to the training data helps transformers learn to attend to the most informative elements, making the analogies learnable.
- The authors report improved generalization to new alphabets when training includes more heterogeneous datasets, with their 3-layer encoder-decoder model outperforming many frontier models.
- The work shows partial generalization to compositions of transformations but limited ability to handle completely novel transformations, and it proposes an algorithmic approximation of the model’s computations.
- Interpretability analyses indicate the model’s behavior can be steered in controlled ways consistent with the proposed algorithm, offering insights into parallels with human analogical reasoning.
Related Articles

Black Hat Asia
AI Business

Amazon CEO takes aim at Nvidia, Intel, Starlink, more in annual shareholder letter
TechCrunch

Why Anthropic’s new model has cybersecurity experts rattled
Reddit r/artificial
Does the AI 2027 paper still hold any legitimacy?
Reddit r/artificial

Why Most Productivity Systems Fail (And What to Do Instead)
Dev.to