Exploring the Limits of Pruning: Task-Specific Neurons, Model Collapse, and Recovery in Task-Specific Large Language Models

arXiv cs.CL / 5/1/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper examines neuron pruning in task-specific LLMs and tests whether all neurons contribute uniformly to specialized performance for math reasoning and code generation.
  • It introduces an activation-based selectivity metric to identify and prune low-contribution neurons, showing that selective pruning consistently beats random pruning at preserving target-task accuracy.
  • Reverse-pruning results indicate that removing only ~10% of the most task-specific neurons can trigger a complete model performance collapse, implying critical task information is concentrated in a small network subset.
  • The study finds a pruning robustness threshold of about 15–20% for 1.5B and 7B models, after which accuracy drops and generation failures rise sharply.
  • Fine-tuning after pruning substantially restores performance across pruning levels, especially for more aggressively pruned models, while pruning reduces parameters, VRAM usage, and improves inference throughput.

Abstract

Neuron pruning is widely used to reduce the computational cost and parameter footprint of large language models, yet it remains unclear whether neurons in task-specific models contribute uniformly to task performance. In this work, we provide empirical evidence for the existence and importance of task-specific neurons through a systematic pruning study on language models specialized for mathematical reasoning and code generation. We introduce an activation-based selectivity metric to identify neurons with low contribution to the target task and prune them while preserving target-task accuracy, and compare selective pruning with random pruning. Selective pruning consistently outperforms random pruning, indicating that activation-based selectivity provides a systematic advantage over random pruning. Reverse pruning experiments further show that removing a small subset of highly task-specific neurons (~10%) causes complete performance collapse, suggesting that there exist task specific neurons and critical task information is concentrated in a small portion of the network. In contrast, selective pruning of less critical neurons (~30% - ~35%) reduces accuracy but still preserves significant performance. We also observed consistent reductions in parameters and runtime VRAM usage, along with improved inference throughput as pruning increases. Experiments on both 1.5B and 7B models reveal a robustness threshold around 15-20% pruning, beyond which accuracy loss and generation failures increase sharply. Fine-tuning substantially recovers performance across pruning levels, particularly for aggressively pruned models. These findings provide empirical evidence of neuron specialization in task-specific language models and offer insights into pruning robustness, model redundancy, and post-pruning recoverability.