[1]
Y. Tian, C. Wang, J. Han, Y. Tang, and K. Han, “PocketLLM: Ultimate Compression of Large Language Models via Meta Networks”, AAAI, vol. 40, no. 39, pp. 33250–33258, Mar. 2026.