Domain-Controlled Prompt Learning
DOI:
https://doi.org/10.1609/aaai.v38i2.27853Keywords:
CV: Language and Vision, CV: Large Vision Models, CV: Multi-modal VisionAbstract
Large pre-trained vision-language models, such as CLIP, have shown remarkable generalization capabilities across various tasks when appropriate text prompts are provided. However, adapting these models to specific domains, like remote sensing images (RSIs), medical images, etc, remains unexplored and challenging. Existing prompt learning methods often lack domain-awareness or domain-transfer mechanisms, leading to suboptimal performance due to the misinterpretation of specific images in natural image patterns. To tackle this dilemma, we proposed a Domain-Controlled Prompt Learning for the specific domains. Specifically, the large-scale specific domain foundation model (LSDM) is first introduced to provide essential specific domain knowledge. Using lightweight neural networks, we transfer this knowledge into domain biases, which control both the visual and language branches to obtain domain-adaptive prompts in a directly incorporating manner. Simultaneously, to overcome the existing overfitting challenge, we propose a novel noisy-adding strategy, without extra trainable parameters, to help the model escape the suboptimal solution in a global domain oscillation manner. Experimental results show our method achieves state-of-the-art performance in specific domain image recognition datasets. Our code is available at https://github.com/caoql98/DCPL.Downloads
Published
2024-03-24
How to Cite
Cao, Q., Xu, Z., Chen, Y., Ma, C., & Yang, X. (2024). Domain-Controlled Prompt Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 38(2), 936-944. https://doi.org/10.1609/aaai.v38i2.27853
Issue
Section
AAAI Technical Track on Computer Vision I