Crosscoding Through Time: Tracking Emergence & Consolidation Of Linguistic Representations Throughout LLM Pretraining
arXiv cs.CL / 5/1/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper addresses a gap in understanding when specific linguistic abilities emerge during LLM pretraining, since conventional benchmarks do not show how such concepts are acquired over time.
- It uses sparse crosscoders to discover and align internal features across different model checkpoints, enabling tracking of linguistic feature evolution during pretraining.
- The authors train crosscoders on open-sourced checkpoint triplets with substantial performance and representation shifts to study changes in learned representations.
- They introduce a new metric, Relative Indirect Effects (RelIE), to identify the training stages when individual features become causally important for task performance.
- Results show that the method can detect phases where features emerge, persist, or discontinue, and the approach is architecture-agnostic and scalable for more interpretable representation-learning analysis.
Related Articles

Why Autonomous Coding Agents Keep Failing — And What Actually Works
Dev.to

Why Enterprise AI Pilots Fail
Dev.to

The PDF Feature Nobody Asked For (That I Use Every Day)
Dev.to

How to Fix OpenClaw Tool Calling Issues
Dev.to

Mistral's new flagship Medium 3.5 folds chat, reasoning, and code into one model
THE DECODER