Transferable Attention for Domain Adaptation

Authors

  • Ximei Wang Tsinghua University
  • Liang Li Tsinghua University
  • Weirui Ye Tsinghua University
  • Mingsheng Long Tsinghua University
  • Jianmin Wang Tsinghua University

DOI:

https://doi.org/10.1609/aaai.v33i01.33015345

Abstract

Recent work in domain adaptation bridges different domains by adversarially learning a domain-invariant representation that cannot be distinguished by a domain discriminator. Existing methods of adversarial domain adaptation mainly align the global images across the source and target domains. However, it is obvious that not all regions of an image are transferable, while forcefully aligning the untransferable regions may lead to negative transfer. Furthermore, some of the images are significantly dissimilar across domains, resulting in weak image-level transferability. To this end, we present Transferable Attention for Domain Adaptation (TADA), focusing our adaptation model on transferable regions or images. We implement two types of complementary transferable attention: transferable local attention generated by multiple region-level domain discriminators to highlight transferable regions, and transferable global attention generated by single image-level domain discriminator to highlight transferable images. Extensive experiments validate that our proposed models exceed state of the art results on standard domain adaptation datasets.

Downloads

Published

2019-07-17

How to Cite

Wang, X., Li, L., Ye, W., Long, M., & Wang, J. (2019). Transferable Attention for Domain Adaptation. Proceedings of the AAAI Conference on Artificial Intelligence, 33(01), 5345-5352. https://doi.org/10.1609/aaai.v33i01.33015345

Issue

Section

AAAI Technical Track: Machine Learning