How LLMs Follow Instructions: Skillful Coordination, Not a Universal Mechanism

arXiv cs.AI / 4/8/2026

💬 OpinionSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • arXivの新規研究は、一般に「instruction tuningが指示追従の汎用メカニズムを与える」と考えられている点を検証し、9種のタスクと3つのinstruction-tunedモデルで診断プロービングを実施しました。
  • 訓練横断の汎用プローブはタスク専用のモデルより性能が低く、表現の共有が限定的であることを示しています。
  • タスク間の転移は弱く、スキル類似性の塊としてクラスタリングされ、因果的アブレーションでも共通表現ではなく疎で非対称な依存関係が観測されました。
  • 複雑性の違いによりレイヤーでタスクが層化し、構造的制約は早期に、意味的タスクは後期に現れ、さらに動的モニタリングによる生成中の整合であって事前計画型の「単一の制約チェック」ではないと結論づけています。

Abstract

Instruction tuning is commonly assumed to endow language models with a domain-general ability to follow instructions, yet the underlying mechanism remains poorly understood. Does instruction-following rely on a universal mechanism or compositional skill deployment? We investigate this through diagnostic probing across nine diverse tasks in three instruction-tuned models. Our analysis provides converging evidence against a universal mechanism. First, general probes trained across all tasks consistently underperform task-specific specialists, indicating limited representational sharing. Second, cross-task transfer is weak and clustered by skill similarity. Third, causal ablation reveals sparse asymmetric dependencies rather than shared representations. Tasks also stratify by complexity across layers, with structural constraints emerging early and semantic tasks emerging late. Finally, temporal analysis shows constraint satisfaction operates as dynamic monitoring during generation rather than pre-generation planning. These findings indicate that instruction-following is better characterized as skillful coordination of diverse linguistic capabilities rather than deployment of a single abstract constraint-checking process.