Prompt-Based Distribution Alignment for Unsupervised Domain Adaptation

Authors

  • Shuanghao Bai Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China
  • Min Zhang Westlake University Institute of Advanced Technology, Westlake Institute for Advanced Study
  • Wanqi Zhou Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China RIKEN AIP
  • Siteng Huang Westlake University Institute of Advanced Technology, Westlake Institute for Advanced Study
  • Zhirong Luan School of Electrical Engineering, Xi’an University of Technology, Xi'an, China
  • Donglin Wang Westlake University Institute of Advanced Technology, Westlake Institute for Advanced Study
  • Badong Chen Institute of Artificial Intelligence and Robotics, Xi'an Jiaotong University, Xi'an, China

DOI:

https://doi.org/10.1609/aaai.v38i2.27830

Keywords:

CV: Multi-modal Vision, ML: Transfer, Domain Adaptation, Multi-Task Learning

Abstract

Recently, despite the unprecedented success of large pre-trained visual-language models (VLMs) on a wide range of downstream tasks, the real-world unsupervised domain adaptation (UDA) problem is still not well explored. Therefore, in this paper, we first experimentally demonstrate that the unsupervised-trained VLMs can significantly reduce the distribution discrepancy between source and target domains, thereby improving the performance of UDA. However, a major challenge for directly deploying such models on downstream UDA tasks is prompt engineering, which requires aligning the domain knowledge of source and target domains, since the performance of UDA is severely influenced by a good domain-invariant representation. We further propose a Prompt-based Distribution Alignment (PDA) method to incorporate the domain knowledge into prompt learning. Specifically, PDA employs a two-branch prompt-tuning paradigm, namely base branch and alignment branch. The base branch focuses on integrating class-related representation into prompts, ensuring discrimination among different classes. To further minimize domain discrepancy, for the alignment branch, we construct feature banks for both the source and target domains and propose image-guided feature tuning (IFT) to make the input attend to feature banks, which effectively integrates self-enhanced and cross-domain features into the model. In this way, these two branches can be mutually promoted to enhance the adaptation of VLMs for UDA. We conduct extensive experiments on three benchmarks to demonstrate that our proposed PDA achieves state-of-the-art performance. The code is available at https://github.com/BaiShuanghao/Prompt-based-Distribution-Alignment.

Published

2024-03-24

How to Cite

Bai, S., Zhang, M., Zhou, W., Huang, S., Luan, Z., Wang, D., & Chen, B. (2024). Prompt-Based Distribution Alignment for Unsupervised Domain Adaptation. Proceedings of the AAAI Conference on Artificial Intelligence, 38(2), 729-737. https://doi.org/10.1609/aaai.v38i2.27830

Issue

Section

AAAI Technical Track on Computer Vision I