🤖 AI Summary
In LLM inference, KV cache memory consumption grows rapidly with sequence length; existing compression methods impose fixed compression ratios, leading to the “Procrustean bed” problem—rigid resource allocation and degraded performance. This paper proposes the first adaptive KV cache compression framework that eliminates manual compression budget specification. It employs Monte Carlo sampling to simulate future queries and dynamically identifies critical key-value pairs via attention importance scoring, enabling fine-grained aggregation and retention. By decoupling compression from pre-defined ratios, the method achieves demand-driven, real-time cache pruning. Evaluated on GSM8K, RULER, and LongBench, it attains 2× memory reduction while preserving or exceeding baseline accuracy.
📝 Abstract
Large language models (LLMs) inference relies heavily on KV-caches to accelerate autoregressive decoding, but the resulting memory footprint grows rapidly with sequence length, posing significant efficiency challenges. Current KV-cache compression methods suffer from a Procrustes' bed problem: they force diverse workloads into fixed compression ratios, leading to suboptimal resource allocation and inference performance. To this end, we present GVote, an adaptive KV-cache compression scheme that eliminates manual budget specification while achieving superior accuracy-efficiency trade-offs. GVote operates on the principle that the important keys are the aggregation of keys required by future queries. The method predicts future query attention demands by Monte-Carlo style sampling potential queries and aggregating selected keys to determine the optimal cache budget without manual specification. Experimental evaluation demonstrates GVote's effectiveness across multiple benchmarks, including GSM8K, RULER and Longbench. Compared to baselines, GVote exhibits 2$ imes$ memory reduction while the accuracy maintains higher or comparable.