Target-Oriented Pretraining Data Selection via Neuron-Activated Graph

arXiv cs.CL / 4/20/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes Neuron-Activated Graph (NAG) Ranking, a training-free and interpretable method to select target-oriented LM pretraining data based on neuron activation patterns in off-the-shelf LLMs.
  • It builds a compact graph of the most influential neurons across layers, then ranks candidate training samples by how similar they are to target examples using NAG similarity.
  • Experiments on six benchmarks show an average 4.9% improvement over random sampling for target-oriented pretraining, and a 5.3% accuracy advantage over state-of-the-art baselines on HellaSwag.
  • The approach also works in a more practical multi-target setting, where the best configuration outperforms two baselines by 1.1% and 4.1%, respectively.
  • The authors include interpretability analyses indicating that removing NAG-selected neurons (only 0.12% of all) causes a 23.5% performance collapse, and limiting NAG to the final layer leads to a 4.1% average drop, suggesting a sparse “functional backbone” for learning target features.

Abstract

Everyday tasks come with a target, and pretraining models around this target is what turns them into experts. In this paper, we study target-oriented language model (LM) pretraining by introducing Neuron-Activated Graph Ranking (NAG-based Ranking), a training-free and interpretable framework for target pretraining data selection. Rather than using black-box representations, our approach directly characterizes each target input by a sparse set of high-impact neurons in any off-the-shelf LLMs. Concretely, we quantify neuron impact and select the most influential neurons across layers into a compact Neuron-Activated Graph (NAG), and rank candidate data by NAG similarity to target examples. We conduct experiments across six benchmarks, where our NAG-based Ranking improves target-oriented pretraining by 4.9% on average over random sampling, and also outperforms state-of-the-art baselines by 5.3% accuracy on HellaSwag. It also remains effective under a more applicable multi-target setting, where our best setup surpasses two baselines by 1.1% and 4.1%, respectively. Furthermore, we provide a comprehensive analysis on why and how our NAG works, e.g., deactivating NAG-selected neurons (only 0.12% of all) causes a 23.5% performance collapse, and restricting NAG to the final layer incurs a 4.1% average drop, indicating that NAG captures a sparse "functional backbone" for learning target features. We release the code at https://github.com/asillycat/NAG.