From Insight to Action: A Novel Framework for Interpretability-Guided Data Selection in Large Language Models

arXiv cs.AI / 4/29/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes Interpretability-Guided Data Selection (IGDS) to turn mechanistic interpretability findings into actionable training data for LLM fine-tuning.
  • IGDS identifies causal task features using methods like frequency recall and interventional filtering, then selects “Feature-Resonant Data” that most strongly activates those features during training.
  • Experiments on mathematical reasoning, summarization, and translation show IGDS improves model performance across Gemma-2, LLaMA-3.1, and Qwen3.
  • On the math task, IGDS beats full-dataset fine-tuning by 17.4% on Gemma-2-2B while using only half the data, outperforming baselines centered on data quality/diversity.
  • The analysis finds a strong positive link between feature amplification and task performance gains, supporting the authors’ core hypothesis.

Abstract

While mechanistic interpretability tools like Sparse Autoencoders (SAEs) can uncover meaningful features within Large Language Models (LLMs), a critical gap remains in transforming these insights into practical actions for model optimization. We bridge this gap with the hypothesis that data selection guided by a model's internal task features is a effective training strategy. Inspired by this, we propose Interpretability-Guided Data Selection (IGDS), a framework that first identifies these causal task features through frequency recall and interventional filtering, then selects ``Feature-Resonant Data'' that maximally activates task features for fine-tuning. We validate IGDS on mathematical reasoning, summarization, and translation tasks within Gemma-2, LLaMA-3.1, and Qwen3 models. Our experiments demonstrate exceptional data efficiency: on the Math task, IGDS surpasses full-dataset fine-tuning by a remarkable 17.4% on Gemma-2-2B while using only 50% of the data, and outperforms established baselines focused on data quality and diversity. Analysis confirms a strong positive correlation between feature amplification and task performance improvement. IGDS thus provides a direct and effective framework to enhance LLMs by leveraging their internal mechanisms, validating our core hypothesis.