Stochastic KV Routing: Enabling Adaptive Depth-Wise Cache Sharing

arXiv cs.LG / 4/28/2026

📰 NewsDeveloper Stack & InfrastructureModels & Research

Key Points

  • The paper addresses the high memory cost of KV (key-value) caching in transformer language model serving by targeting optimization along the depth dimension rather than only temporal-axis compression/eviction.
  • It argues that maintaining a full KV cache for every layer can be redundant, but existing cross-layer KV sharing approaches often reduce throughput or increase time-to-first-token.
  • The authors propose “stochastic KV routing,” where during training each layer randomly attends to either its own KV states or those from a preceding layer via random cross-layer attention.
  • Experiments show that this stochastic training strategy enables depth-wise KV cache sharing across multiple model families during pre-training or fine-tuning, reducing cache memory footprint with no information loss in the proposed setup.
  • In larger, data-constrained scenarios, the method may act like a regularization effect, often preserving or improving performance while substantially lowering KV cache memory usage.

Abstract

Serving transformer language models with high throughput requires caching Key-Values (KVs) to avoid redundant computation during autoregressive generation. The memory footprint of KV caching is significant and heavily impacts serving costs. This work proposes to lessen these memory requirements. While recent work has largely addressed KV cache reduction via compression and eviction along the temporal axis, we argue that the \emph{depth} dimension offers an orthogonal and robust avenue for optimization. Although prior research suggests that a full cache for every layer is redundant, implementing cross-layer cache sharing remains a practical challenge; existing methods typically suffer from reduced throughput or increased time-to-first-token. In this paper, we demonstrate that dropping a layer's cache offers efficient optimization without information loss. We propose a simple training approach: random cross-layer attention. During training, layers randomly choose to attend either to their own KV states or those of a preceding layer. This stochastic process adapts the model to be robust to various depth-wise cache sharing strategies, ensuring flexibility for unknown hardware constraints at deployment time. Our evaluations show that applying this scheme during pre-training or fine-tuning enables depth-wise cache sharing for various model families. Furthermore, for larger models in data-constrained settings, this approach is suggestive of a regularization-like effect, frequently preserving or improving performance while significantly reducing the cache's memory footprint.