[1]
Wang, Q. et al. 2026. Symphony-MoE: Harmonizing Disparate Pre-trained Models into a Coherent Mixture-of-Experts. Proceedings of the AAAI Conference on Artificial Intelligence. 40, 31 (Mar. 2026), 26407–26415. DOI:https://doi.org/10.1609/aaai.v40i31.39847.