Shared Lexical Task Representations Explain Behavioral Variability In LLMs
arXiv cs.AI / 4/27/2026
💬 OpinionIdeas & Deep AnalysisModels & Research
Key Points
- The paper examines why LLM performance varies unpredictably with prompt wording by comparing instruction-based prompting versus example-based (few-shot) prompting styles.
- It finds that, although overall behavior can change greatly across prompts, the model uses shared underlying mechanisms for the same task.
- The authors identify “lexical task heads,” task-specific attention heads whose outputs explicitly reflect the task and help trigger subsequent answer generation across different prompting styles.
- The degree to which these lexical task heads are activated helps explain prompt-to-prompt behavioral variability, and some failures are attributed to competing task representations weakening the target signal.
Related Articles

Subagents: The Building Block of Agentic AI
Dev.to

DeepSeek-V4 Models Could Change Global AI Race
AI Business

Got OpenAI's privacy filter model running on-device via ExecuTorch
Reddit r/LocalLLaMA

The Agent-Skill Illusion: Why Prompt-Based Control Fails in Multi-Agent Business Consulting Systems
Dev.to

We Built a Voice AI Receptionist in 8 Weeks — Every Decision We Made and Why
Dev.to