AutoCompress: Critical Layer Isolation for Efficient Transformer Compression

arXiv cs.LG / 4/28/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • AutoCompress introduces a transformer compression approach based on an empirical observation that Layer 0 holds disproportionately high task-critical information compared with other layers.
  • The method’s Critical Layer Isolation (CLI) architecture preserves Layer 0 at full dimensionality, compresses intermediate layers via a learned bottleneck, and restores full dimensionality at the final layer.
  • On GPT-2 Medium (354.8M parameters), CLI-GPT2 reaches 204.5 perplexity on WikiText-103 using 143.8M parameters, achieving a 2.47× compression ratio and a 59.5% parameter reduction.
  • Ablation results show that a uniform bottleneck of similar size performs far worse (571.8 perplexity), indicating the key benefit comes from isolating/protecting Layer 0 rather than merely shrinking the model.
  • The authors provide public code and checkpoints for reproducing and extending the approach.

Abstract

We present AutoCompress, a transformer compression method motivated by an empirical finding: in small transformers, Layer 0 carries disproportionately high task-critical information, with an NTK-based importance score of 3.6 compared to a maximum of 0.054 for all other layers -- a gap of over 60x. Based on this finding, we propose Critical Layer Isolation (CLI), an architecture that protects Layer 0 at full dimensionality, compresses all intermediate layers through a learned bottleneck, and restores the full dimension at the final layer. Applied to GPT-2 Medium (354.8M parameters), CLI-GPT2 achieves 204.5 perplexity on WikiText-103 with only 143.8M parameters -- a 2.47x compression ratio and 59.5% parameter reduction. Crucially, an ablation study demonstrates that a uniform bottleneck baseline of comparable size achieves only 571.8 perplexity under identical training conditions, confirming that the architectural decision to protect Layer 0 -- rather than simply reducing model size -- is the primary driver of performance. Code and checkpoints are publicly available.