Doing More With Less: Revisiting the Effectiveness of LLM Pruning for Test-Time Scaling

arXiv cs.LG / 4/29/2026

💬 OpinionIdeas & Deep AnalysisModels & Research

Key Points

  • The paper revisits prior claims that structured pruning harms test-time scaling (TTS) reasoning performance in large language models.
  • Experiments on two reasoning-focused LLMs (s1.1-7B and Qwen3-8B) across four benchmarks show that unstructured pruning can improve TTS performance versus structured pruning.
  • In some cases, carefully applied unstructured pruning even outperforms the unpruned full-weight models while retaining better reasoning under increased test-time compute.
  • The authors further analyze how different layer-wise sparsity allocation strategies affect unstructured pruning outcomes, highlighting sparsity allocation as a key design parameter.
  • Overall, the results challenge the conventional assumption that pruning invariably degrades TTS reasoning effectiveness and suggest pruning can be leveraged to make TTS more effective.

Abstract

While current Large Language Models (LLMs) exhibit remarkable reasoning capabilities through test-time compute scaling (TTS), their massive parameter counts and high inference costs have motivated the development of pruning methods that can reduce model size without sacrificing performance. However, specific to reasoning LLMs, prior work has shown that structured pruning (methods which removes entire set of layer blocks), significantly degrades TTS reasoning performance. In this work, we revisit this assumption and instead investigate whether unstructured pruning (methods that carefully remove only certain redundant/detrimental weights) exhibits similar limitations. Surprisingly, our extensive experiments across four reasoning benchmarks on two reasoning LLMs: s1.1-7B and Qwen3-8B, consistently show that unstructured pruning augments TTS performance compared to structured pruning, and at times can even outperform the unpruned full-weight LLMs. Furthermore, we also empirically study the impact of different layer-wise sparsity allocation strategies, which are an important parametric choice for instantiating unstructured pruning methods. These findings challenge the conventional notion that pruning always reduces TTS performance and in fact, suggest that carefully undertaken pruning can improve TTS effectiveness even further.