Sparser, Faster, Lighter Transformer Language Models

arXiv cs.LG / 3/25/2026

💬 OpinionDeveloper Stack & InfrastructureIdeas & Deep AnalysisModels & Research

Key Points

  • The paper proposes reducing the computational cost of autoregressive LLMs by exploiting unstructured sparsity specifically in feedforward layers, which dominate parameters and FLOPs.
  • It introduces a new sparse “packing” format plus CUDA kernels intended to plug into modern GPU execution pipelines for efficient sparse computation in both inference and training.
  • The authors report that L1 regularization can induce over 99% sparsity with negligible impact on downstream model performance, supported by a quantitative sparsity study.
  • With the proposed sparsity and kernels, they claim substantial improvements in throughput, energy efficiency, and memory usage, with benefits that grow as model scale increases.
  • The work plans to release code and kernels under an open-source license to encourage adoption and further research into sparsity as an efficiency lever for foundation models.

Abstract

Scaling autoregressive large language models (LLMs) has driven unprecedented progress but comes with vast computational costs. In this work, we tackle these costs by leveraging unstructured sparsity within an LLM's feedforward layers, the components accounting for most of the model parameters and execution FLOPs. To achieve this, we introduce a new sparse packing format and a set of CUDA kernels designed to seamlessly integrate with the optimized execution pipelines of modern GPUs, enabling efficient sparse computation during LLM inference and training. To substantiate our gains, we provide a quantitative study of LLM sparsity, demonstrating that simple L1 regularization can induce over 99% sparsity with negligible impact on downstream performance. When paired with our kernels, we show that these sparsity levels translate into substantial throughput, energy efficiency, and memory usage benefits that increase with model scale. We will release all code and kernels under an open-source license to promote adoption and accelerate research toward establishing sparsity as a practical axis for improving the efficiency and scalability of modern foundation models.