Localizing Task Recognition and Task Learning in In-Context Learning via Attention Head Analysis
arXiv cs.CL / 4/13/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper studies in-context learning (ICL) by combining attention-head mechanistic analysis with a holistic decomposition into Task Recognition (TR) and Task Learning (TL).
- It introduces Task Subspace Logit Attribution (TSLA) to identify which attention heads specialize in TR versus TL, and shows these heads independently and effectively represent the corresponding ICL components.
- Correlation, ablation, and input-perturbation experiments provide evidence that TR and TL heads play distinct yet complementary roles in performing ICL.
- Steering experiments using geometric analysis of hidden states suggest TR heads align hidden representations with a task subspace, while TL heads rotate those representations within the subspace toward the correct label.
- The authors argue their framework unifies earlier ICL mechanism findings (e.g., induction heads and task vectors) with the TR–TL attention-head perspective for a more interpretable account.
Related Articles

Why Fashion Trend Prediction Isn’t Enough Without Generative AI
Dev.to

Chatbot vs Voicebot: The Real Business Decision Nobody Talks About
Dev.to

วิธีใช้ AI ทำ SEO ให้เว็บติดอันดับ Google (2026)
Dev.to

Free AI Tools With No Message Limits — The Definitive List (2026)
Dev.to

Why Domain Knowledge Is Critical in Healthcare Machine Learning
Dev.to