Agile Multi-Source-Free Domain Adaptation

Authors

  • Xinyao Li University of Electronic Science and Technology of China
  • Jingjing Li University of Electronic Science and Technology of China (UESTC) Shenzhen Institute for Advanced Study, UESTC
  • Fengling Li University of Technology Sydney
  • Lei Zhu School of Electronic and Information Engineering, Tongji University
  • Ke Lu University of Electronic Science and Technology of China

DOI:

https://doi.org/10.1609/aaai.v38i12.29272

Keywords:

ML: Transfer, Domain Adaptation, Multi-Task Learning, ML: Ensemble Methods, ML: Unsupervised & Self-Supervised Learning

Abstract

Efficiently utilizing rich knowledge in pretrained models has become a critical topic in the era of large models. This work focuses on adaptively utilize knowledge from multiple source-pretrained models to an unlabeled target domain without accessing the source data. Despite being a practically useful setting, existing methods require extensive parameter tuning over each source model, which is computationally expensive when facing abundant source domains or larger source models. To address this challenge, we propose a novel approach which is free of the parameter tuning over source backbones. Our technical contribution lies in the Bi-level ATtention ENsemble (Bi-ATEN) module, which learns both intra-domain weights and inter-domain ensemble weights to achieve a fine balance between instance specificity and domain consistency. By slightly tuning source bottlenecks, we achieve comparable or even superior performance on a challenging benchmark DomainNet with less than 3% trained parameters and 8 times of throughput compared with SOTA method. Furthermore, with minor modifications, the proposed module can be easily equipped to existing methods and gain more than 4% performance boost. Code is available at https://github.com/TL-UESTC/Bi-ATEN.

Published

2024-03-24

How to Cite

Li, X., Li, J., Li, F., Zhu, L., & Lu, K. (2024). Agile Multi-Source-Free Domain Adaptation. Proceedings of the AAAI Conference on Artificial Intelligence, 38(12), 13673-13681. https://doi.org/10.1609/aaai.v38i12.29272

Issue

Section

AAAI Technical Track on Machine Learning III