ScoutAttention: Efficient KV Cache Offloading via Layer-Ahead CPU Pre-computation for LLM Inference

arXiv cs.LG / 3/31/2026

📰 NewsIdeas & Deep AnalysisModels & Research

Key Points

  • ScoutAttention is a new KV-cache offloading framework designed to address GPU memory limits during long-context LLM inference, where KV cache size restricts decode batch sizes.
  • The approach uses collaborative GPU-CPU, block-wise sparse attention to reduce CPU load and mitigate the GPU underutilization caused by I/O latency or heavy CPU computation in prior offloading methods.
  • A key contribution is a layer-ahead CPU pre-computation algorithm, allowing the CPU to start attention computation one layer early, with asynchronous periodic recall to keep CPU work minimal.
  • Experiments on the proposed method report accuracy within 2.4% of a baseline and a 2.1× speedup over existing offloading techniques, while maintaining usable long-context performance.

Abstract

Large language models encounter critical GPU memory capacity constraints during long-context inference, where KV cache memory consumption severely limits decode batch sizes. While existing research has explored offloading KV cache to DRAM, these approaches either demand frequent GPU-CPU data transfers or impose extensive CPU computation requirements, resulting in poor GPU utilization as the system waits for I/O operations or CPU processing to complete. We propose ScoutAttention, a novel KV cache offloading framework that accelerates LLM inference through collaborative GPU-CPU attention computation. To prevent CPU computation from bottlenecking the system, ScoutAttention introduces GPU-CPU collaborative block-wise sparse attention that significantly reduces CPU load. Unlike conventional parallel computing approaches, our framework features a novel layer-ahead CPU pre-computation algorithm, enabling the CPU to initiate attention computation one layer in advance, complemented by asynchronous periodic recall mechanisms to maintain minimal CPU compute load. Experimental results demonstrate that ScoutAttention maintains accuracy within 2.4% of baseline while achieving 2.1x speedup compared to existing offloading methods.