To See the Unseen: on the Generalization Ability of Transformers in Symbolic Reasoning
arXiv cs.AI / 4/25/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper studies how decoder-only transformer models can generalize in abstract symbolic reasoning tasks, especially propositional logic problems given in-context examples.
- It explains prior failures with unseen variable names by showing a “representational collapse,” where the unembedding vectors for unseen tokens converge to nearly the same representation during training.
- The collapse makes it hard for models to distinguish between different unseen variables, offering a mechanistic rationale for why heuristic methods like “active forgetting” can help by periodically resetting token (un)embeddings.
- The authors propose a combined approach—small architectural changes to improve copying, more diverse training data, and strategies such as freezing or resetting (un)embeddings—that improves generalization to unseen tokens, supported by extensive controlled experiments.
- They also find evidence of similar (un)embedding collapse in open-weight Gemma 3 family models, noting that correlated embeddings among reserved unused tokens can be a weak starting point for fine-tuning.
Related Articles
Navigating WooCommerce AI Integrations: Lessons for Agencies & Developers from a Bluehost Conflict
Dev.to

One Day in Shenzhen, Seen Through an AI's Eyes
Dev.to

Underwhelming or underrated? DeepSeek V4 shows “impressive” gains
SCMP Tech

Claude Code: Hooks, Subagents, and Skills — Complete Guide
Dev.to

Finding the Gold: An AI Framework for Highlight Detection
Dev.to