Tian, Y. (2026) “KeepKV: Achieving Periodic Lossless KV Cache Compression for Efficient LLM Inference”, Proceedings of the AAAI Conference on Artificial Intelligence, 40(39), pp. 33259–33267. doi: 10.1609/aaai.v40i39.40611.