Beyond Token Eviction: Mixed-Dimension Budget Allocation for Efficient KV Cache Compression

arXiv cs.LG / 3/24/2026

📰 NewsSignals & Early TrendsIdeas & Deep AnalysisModels & Research

Key Points

  • The paper addresses KV-cache memory growth in transformer inference, which limits long-context deployment, by moving beyond token eviction’s coarse zero-or-full dimensionality reduction.
  • It introduces MixedDimKV, which assigns KV cache dimensions to tokens at a finer granularity, and MixedDimKV-H, which additionally uses head-level importance information.
  • Experiments on long-context benchmarks indicate MixedDimKV outperforms prior KV cache compression methods that do not incorporate head importance profiling.
  • With the same head-level importance signals, MixedDimKV-H consistently beats HeadKV, achieving performance close to full attention on LongBench using only 6.25% of the KV cache.
  • In the Needle-in-a-Haystack evaluation, the method preserves 100% accuracy at 50K context length while reducing KV cache usage to as low as 0.26%.

Abstract

Key-value (KV) caching is widely used to accelerate transformer inference, but its memory cost grows linearly with input length, limiting long-context deployment. Existing token eviction methods reduce memory by discarding less important tokens, which can be viewed as a coarse form of dimensionality reduction that assigns each token either zero or full dimension. We propose MixedDimKV, a mixed-dimension KV cache compression method that allocates dimensions to tokens at a more granular level, and MixedDimKV-H, which further integrates head-level importance information. Experiments on long-context benchmarks show that MixedDimKV outperforms prior KV cache compression methods that do not rely on head-level importance profiling. When equipped with the same head-level importance information, MixedDimKV-H consistently outperforms HeadKV. Notably, our approach achieves comparable performance to full attention on LongBench with only 6.25% of the KV cache. Furthermore, in the Needle-in-a-Haystack test, our solution maintains 100% accuracy at a 50K context length while using as little as 0.26% of the cache.