FedDiv: Collaborative Noise Filtering for Federated Learning with Noisy Labels

Authors

  • Jichang Li School of Computer Science and Engineering, Research Institute of Sun Yat-sen University in Shenzhen, Sun Yat-sen University, Guangzhou, China Department of Computer Science, The University of Hong Kong, Hong Kong
  • Guanbin Li School of Computer Science and Engineering, Research Institute of Sun Yat-sen University in Shenzhen, Sun Yat-sen University, Guangzhou, China Guangdong Province Key Laboratory of Information Security Technology
  • Hui Cheng UM-SJTU Joint Institute, Shanghai Jiao Tong University, Shanghai, China
  • Zicheng Liao Department of Computer Science, The University of Hong Kong, Hong Kong Zhejiang University
  • Yizhou Yu Department of Computer Science, The University of Hong Kong, Hong Kong

DOI:

https://doi.org/10.1609/aaai.v38i4.28095

Keywords:

CV: Object Detection & Categorization, ML: Distributed Machine Learning & Federated Learning, ML: Semi-Supervised Learning, ML: Unsupervised & Self-Supervised Learning

Abstract

Federated Learning with Noisy Labels (F-LNL) aims at seeking an optimal server model via collaborative distributed learning by aggregating multiple client models trained with local noisy or clean samples. On the basis of a federated learning framework, recent advances primarily adopt label noise filtering to separate clean samples from noisy ones on each client, thereby mitigating the negative impact of label noise. However, these prior methods do not learn noise filters by exploiting knowledge across all clients, leading to sub-optimal and inferior noise filtering performance and thus damaging training stability. In this paper, we present FedDiv to tackle the challenges of F-LNL. Specifically, we propose a global noise filter called Federated Noise Filter for effectively identifying samples with noisy labels on every client, thereby raising stability during local training sessions. Without sacrificing data privacy, this is achieved by modeling the global distribution of label noise across all clients. Then, in an effort to make the global model achieve higher performance, we introduce a Predictive Consistency based Sampler to identify more credible local data for local model training, thus preventing noise memorization and further boosting the training stability. Extensive experiments on CIFAR-10, CIFAR-100, and Clothing1M demonstrate that FedDiv achieves superior performance over state-of-the-art F-LNL methods under different label noise settings for both IID and non-IID data partitions. Source code is publicly available at https://github.com/lijichang/FLNL-FedDiv.

Published

2024-03-24

How to Cite

Li, J., Li, G., Cheng, H., Liao, Z., & Yu, Y. (2024). FedDiv: Collaborative Noise Filtering for Federated Learning with Noisy Labels. Proceedings of the AAAI Conference on Artificial Intelligence, 38(4), 3118-3126. https://doi.org/10.1609/aaai.v38i4.28095

Issue

Section

AAAI Technical Track on Computer Vision III