As Language Models Scale, Low-order Linear Depth Dynamics Emerge
arXiv cs.LG / 3/16/2026
📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research
Key Points
- A 32-dimensional linear surrogate can accurately reproduce the layerwise sensitivity profile of GPT-2-large across multiple tasks such as toxicity, irony, hate speech, and sentiment.
- The surrogate reveals how the final output shifts when small additive injections are made at each layer, enabling precise, interpretable analysis of depth dynamics.
- The authors uncover a scaling principle: for a fixed-order surrogate, agreement with the full model improves monotonically as model size increases across the GPT-2 family.
- The linear surrogate enables principled multi-layer interventions that use less energy than standard heuristic schedules when applied to the full model.
- Together, the results suggest that as language models scale, low-order linear depth dynamics emerge, providing a systems-theoretic basis for analysis and control.
Related Articles

Interactive Web Visualization of GPT-2
Reddit r/artificial
Stop Treating AI Interview Fraud Like a Proctoring Problem
Dev.to
[R] Causal self-attention as a probabilistic model over embeddings
Reddit r/MachineLearning
The 5 software development trends that actually matter in 2026 (and what they mean for your startup)
Dev.to
InVideo AI Review: Fast Finished
Dev.to