1.
Liu J, He Y, Ke W, Wang P, Shang Z, Li G, et al. Balanced Knowledge Distillation for Large Language Models with Mix-of-Experts. AAAI [Internet]. 2026 Mar. 14 [cited 2026 May 15];40(28):23694-702. Available from: https://ojs.aaai.org/index.php/AAAI/article/view/39543