Task-Adaptive Prompted Transformer for Cross-Domain Few-Shot Learning
DOI:
https://doi.org/10.1609/aaai.v38i6.28416Keywords:
CV: Representation Learning for Vision, ML: Transfer, Domain Adaptation, Multi-Task Learning, CV: Learning & Optimization for CVAbstract
Cross-Domain Few-Shot Learning (CD-FSL) aims at recognizing samples in novel classes from unseen domains that are vastly different from training classes, with few labeled samples. However, the large domain gap between training and novel classes makes previous FSL methods perform poorly. To address this issue, we propose MetaPrompt, a Task-adaptive Prompted Transformer model for CD-FSL, by jointly exploiting prompt learning and the parameter generation framework. The proposed MetaPrompt enjoys several merits. First, a task-conditioned prompt generator is established upon attention mechanisms. It can flexibly produce a task-adaptive prompt with arbitrary length for unseen tasks, by selectively gathering task characteristics from the contextualized support embeddings. Second, the task-adaptive prompt is attached to Vision Transformer to facilitate fast task adaptation, steering the task-agnostic representation to incorporate task knowledge. To our best knowledge, this is the first work to exploit a prompt-based parameter generation mechanism for CD-FSL. Extensive experimental results on the Meta-Dataset benchmark demonstrate that our method achieves superior results against state-of-the-art methods.Downloads
Published
2024-03-24
How to Cite
Wu, J., Liu, X., Yin, X., Zhang, T., & Zhang, Y. (2024). Task-Adaptive Prompted Transformer for Cross-Domain Few-Shot Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 38(6), 6012-6020. https://doi.org/10.1609/aaai.v38i6.28416
Issue
Section
AAAI Technical Track on Computer Vision V