Task-Adaptive Prompted Transformer for Cross-Domain Few-Shot Learning

Authors

  • Jiamin Wu Deep Space Exploration Laboratory/School of Information Science and Technology, University of Science and Technology of China
  • Xin Liu Deep Space Exploration Laboratory/School of Information Science and Technology, University of Science and Technology of China
  • Xiaotian Yin Deep Space Exploration Laboratory/School of Information Science and Technology, University of Science and Technology of China
  • Tianzhu Zhang Deep Space Exploration Laboratory/School of Information Science and Technology, University of Science and Technology of China
  • Yongdong Zhang Deep Space Exploration Laboratory/School of Information Science and Technology, University of Science and Technology of China

DOI:

https://doi.org/10.1609/aaai.v38i6.28416

Keywords:

CV: Representation Learning for Vision, ML: Transfer, Domain Adaptation, Multi-Task Learning, CV: Learning & Optimization for CV

Abstract

Cross-Domain Few-Shot Learning (CD-FSL) aims at recognizing samples in novel classes from unseen domains that are vastly different from training classes, with few labeled samples. However, the large domain gap between training and novel classes makes previous FSL methods perform poorly. To address this issue, we propose MetaPrompt, a Task-adaptive Prompted Transformer model for CD-FSL, by jointly exploiting prompt learning and the parameter generation framework. The proposed MetaPrompt enjoys several merits. First, a task-conditioned prompt generator is established upon attention mechanisms. It can flexibly produce a task-adaptive prompt with arbitrary length for unseen tasks, by selectively gathering task characteristics from the contextualized support embeddings. Second, the task-adaptive prompt is attached to Vision Transformer to facilitate fast task adaptation, steering the task-agnostic representation to incorporate task knowledge. To our best knowledge, this is the first work to exploit a prompt-based parameter generation mechanism for CD-FSL. Extensive experimental results on the Meta-Dataset benchmark demonstrate that our method achieves superior results against state-of-the-art methods.

Published

2024-03-24

How to Cite

Wu, J., Liu, X., Yin, X., Zhang, T., & Zhang, Y. (2024). Task-Adaptive Prompted Transformer for Cross-Domain Few-Shot Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 38(6), 6012-6020. https://doi.org/10.1609/aaai.v38i6.28416

Issue

Section

AAAI Technical Track on Computer Vision V