Task Vectors, Learned Not Extracted: Performance Gains and Mechanistic Insight
arXiv cs.CL / 4/13/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper argues that in-context learning task “demonstrations” can be represented by task vectors (TVs), but prior extraction methods were cumbersome and didn’t explain how TVs actually affect computation.
- It introduces Learned Task Vectors (LTVs) trained directly rather than extracted from hidden states/outputs, reporting improved accuracy, flexibility across layers/positions, and compatibility with in-context learning prompts.
- Through mechanistic analysis, the authors show TVs influence predictions mainly via attention-head OV (output projection) circuits, with a small set of “key heads” driving most of the effect.
- They further find that despite transformer nonlinearities, TV propagation is largely linear across the network, where early TVs rotate into task-relevant subspaces to improve relevant label logits, and later TVs mainly adjust (scale) magnitude.
Related Articles

Why Fashion Trend Prediction Isn’t Enough Without Generative AI
Dev.to

Chatbot vs Voicebot: The Real Business Decision Nobody Talks About
Dev.to

วิธีใช้ AI ทำ SEO ให้เว็บติดอันดับ Google (2026)
Dev.to

Free AI Tools With No Message Limits — The Definitive List (2026)
Dev.to

Why Domain Knowledge Is Critical in Healthcare Machine Learning
Dev.to