Long, L. (2026) “SlimInfer: Accelerating Long-Context LLM Inference via Dynamic Token Pruning”, Proceedings of the AAAI Conference on Artificial Intelligence, 40(38), pp. 32284–32292. doi: 10.1609/aaai.v40i38.40502.