SparK: Query-Aware Unstructured Sparsity with Recoverable KV Cache Channel Pruning
DOI:
https://doi.org/10.1609/aaai.v40i38.40466Abstract
Long-context inference in large language models (LLMs) is increasingly constrained by the KV cache bottleneck: memory usage grows linearly with sequence length, while attention computation scales quadratically. Existing approaches address this issue by compressing the KV cache along the temporal axis through strategies such as token eviction or merging to reduce memory and computational overhead. However, these methods often neglect fine-grained importance variations across feature dimensions (i.e., the channel axis), thereby limiting their ability to effectively balance efficiency and model accuracy. In reality, we observe that channel saliency varies dramatically across both queries and positions: certain feature channels carry near-zero information for a given query, while others spike in relevance. To address this oversight, we propose SPARK, a training-free plug-and-play method that applies unstructured sparsity by pruning KV at the channel level, while dynamically restoring the pruned entries during attention score computation. Notably, our approach is orthogonal to existing KV compression and quantization techniques, making it compatible for integration with them to achieve further acceleration. By reducing channel-level redundancy, SPARK enables processing of longer sequences within the same memory budget. For sequences of equal length, SPARK not only preserves or improves model accuracy but also reduces KV cache storage by over 30% compared to eviction-based methods. Furthermore, even in an aggressive pruning ratio of 80%, SPARK maintains performance with less degradation than 5% compared to the based eviction method, demonstrating robustness and effectiveness. Our code will be available at \url{https://github.com/AMD-AIG-AIMA/AMD-Spark}.Published
2026-03-14
How to Cite
Liao, H., Xu, Y., He, S., Li, G., Yin, X., Li, D., … Liu, K. (2026). SparK: Query-Aware Unstructured Sparsity with Recoverable KV Cache Channel Pruning. Proceedings of the AAAI Conference on Artificial Intelligence, 40(38), 31961–31969. https://doi.org/10.1609/aaai.v40i38.40466
Issue
Section
AAAI Technical Track on Natural Language Processing III