SWIFT: A Scalable Lightweight Infrastructure for Fine-Tuning

Authors

  • Yuze Zhao Alibaba Group
  • Jintao Huang Alibaba Group
  • Jinghan Hu Alibaba Group
  • Xingjun Wang Alibaba Group
  • Yunlin Mao Alibaba Group
  • Daoze Zhang Alibaba Group
  • Zeyinzi Jiang Alibaba Group
  • Zhikai Wu Alibaba Group
  • Baole Ai Alibaba Group
  • Ang Wang Alibaba Group
  • Wenmeng Zhou Alibaba Group
  • Yingda Chen Alibaba Group

DOI:

https://doi.org/10.1609/aaai.v39i28.35383

Abstract

Recent development in Large Language Models (LLMs) and Multi-modal Large Language Models (MLLMs) have achieved superior performance and generalization capabilities, covered extensive areas of traditional tasks. However, existing large model training frameworks support only a limited number of models and techniques, particularly lacking in support for new models, which makes fine-tuning LLMs challenging for most developers. Therefore, we develop SWIFT, a customizable one-stop infrastructure for large models. With support of over 350+ LLMs and 80+ MLLMs, SWIFT stands as the open-source framework that provide the most comprehensive support for fine-tuning large models. In particular, it is the first training framework that provides systematic support for MLLMs. Moreover, SWIFT integrates post-training processes such as inference, evaluation, and quantization, to facilitate fast adoptions of large models in various application scenarios, offering helpful utilities like benchmark comparisons among different training techniques.

Published

2025-04-11

How to Cite

Zhao, Y., Huang, J., Hu, J., Wang, X., Mao, Y., Zhang, D., Jiang, Z., Wu, Z., Ai, B., Wang, A., Zhou, W., & Chen, Y. (2025). SWIFT: A Scalable Lightweight Infrastructure for Fine-Tuning. Proceedings of the AAAI Conference on Artificial Intelligence, 39(28), 29733-29735. https://doi.org/10.1609/aaai.v39i28.35383