@article{He_Chen_Yang_Zhang_Zeng_2022, title={Class-Wise Adaptive Self Distillation for Federated Learning on Non-IID Data (Student Abstract)}, volume={36}, url={https://ojs.aaai.org/index.php/AAAI/article/view/21620}, DOI={10.1609/aaai.v36i11.21620}, abstractNote={Federated learning (FL) enables multiple clients to collaboratively train a globally generalized model while keeping local data decentralized. A key challenge in FL is to handle the heterogeneity of data distributions among clients. The local model will shift the global feature when fitting local data, which results in forgetting the global knowledge. Following the idea of knowledge distillation, the global model’s prediction can be utilized to help local models preserve the global knowledge in FL. However, when the global model hasn’t converged completely, its predictions tend to be less reliable on certain classes, which may results in distillation’s misleading of local models. In this paper, we propose a class-wise adaptive self distillation (FedCAD) mechanism to ameliorate this problem. We design class-wise adaptive terms to soften the influence of distillation loss according to the global model’s performance on each class and therefore avoid the misleading. Experiments show that our method outperforms other state-of-the-art FL algorithms on benchmark datasets.}, number={11}, journal={Proceedings of the AAAI Conference on Artificial Intelligence}, author={He, Yuting and Chen, Yiqiang and Yang, Xiaodong and Zhang, Yingwei and Zeng, Bixiao}, year={2022}, month={Jun.}, pages={12967-12968} }