Tian, Yuxuan, et al. “KeepKV: Achieving Periodic Lossless KV Cache Compression for Efficient LLM Inference”. Proceedings of the AAAI Conference on Artificial Intelligence, vol. 40, no. 39, Mar. 2026, pp. 33259-67, doi:10.1609/aaai.v40i39.40611.