Multi-Modal Style Transfer-based Prompt Tuning for Efficient Federated Domain Generalization

Authors

  • Yuliang Chen School of Computer Science, Shanghai Jiao Tong University Shanghai Key Laboratory of Integrated Administration Technologies for Information Security
  • Xi Lin School of Computer Science, Shanghai Jiao Tong University Shanghai Key Laboratory of Integrated Administration Technologies for Information Security
  • Jun Wu School of Computer Science, Shanghai Jiao Tong University Shanghai Key Laboratory of Integrated Administration Technologies for Information Security
  • Xiangrui Cai College of Computer Science, Nankai University
  • Qiaolun Zhang Department of Electronics, Information and Bioengineering, Polytechnic Institute of Milan
  • Xichun Fan New York University Shanghai
  • Jiapeng Xu School of Computer Science, Shanghai Jiao Tong University Shanghai Key Laboratory of Integrated Administration Technologies for Information Security
  • Xiu Su Big Data Institute, Central South University

DOI:

https://doi.org/10.1609/aaai.v40i25.39177

Abstract

Federated Domain Generalization (FDG) aims to collaboratively train a global model across distributed clients that can generalize well on unseen domains. However, existing FDG methods typically struggle with cross-client data heterogeneity and incur significant communication and computation overhead. To address these challenges, this paper presents a new FDG framework, dubbed FaST-PT, which facilitates local feature augmentation and efficient unseen domain adaptation in a distributed manner. First, we propose a lightweight Multi-Modal Style Transfer (MST) method to transform image embedding under text supervision, which could expand the training data distribution and mitigate domain shift. We then design a dual-prompt module that decomposes the prompt into global and domain prompts. Specifically, global prompts capture general knowledge from augmented embedding across clients, while domain prompts capture domain-specific knowledge from local data. Besides, Domain-aware Prompt Generation (DPG) is introduced to adaptively generate suitable prompts for each sample, which facilitates unseen domain adaptation through knowledge fusion. Extensive experiments on four cross-domain benchmark datasets, e.g., PACS and DomainNet, demonstrate the superior performance of FaST-PT over SOTA FDG methods such as FedDG-GA and DiPrompt. Ablation studies further validate the effectiveness and efficiency of FaST-PT.

Downloads

Published

2026-03-14

How to Cite

Chen, Y., Lin, X., Wu, J., Cai, X., Zhang, Q., Fan, X., … Su, X. (2026). Multi-Modal Style Transfer-based Prompt Tuning for Efficient Federated Domain Generalization. Proceedings of the AAAI Conference on Artificial Intelligence, 40(25), 20427–20435. https://doi.org/10.1609/aaai.v40i25.39177

Issue

Section

AAAI Technical Track on Machine Learning II