Robust Learning from Noisily Labeled Long-Tailed Data via Fairness Regularizer

Authors

  • Jiaheng Wei The Hong Kong University of Science and Technology
  • Zhaowei Zhu D5Data
  • Gang Niu RIKEN
  • Tongliang Liu The University of Sydney
  • Sijia Liu Michigan State University
  • Masashi Sugiyama RIKEN The University of Tokyo
  • Yang Liu University of California, Santa Cruz

DOI:

https://doi.org/10.1609/aaai.v40i42.40899

Abstract

Both long-tailed and noisily labeled data frequently appear in real-world applications and impose significant challenges for learning. Most prior works treat either problem in an isolated way and do not explicitly consider the coupling effects of the two. Our empirical observation reveals that such solutions fail to consistently improve the learning when the dataset is long-tailed with label noise. Moreover, with the presence of label noise, existing methods do not observe universal improvements across different sub-populations; in other words, some sub-populations enjoyed the benefits of improved accuracy at the cost of hurting others. Based on these observations, we introduce the Fairness Regularizer (FR), inspired by regularizing the performance gap between any two sub-populations. We show that the introduced fairness regularizer improves the performances of sub-populations on the tail and the overall learning performance. Extensive experiments demonstrate the effectiveness of the proposed solution when complemented with certain existing popular robust or class-balanced methods.

Published

2026-03-14

How to Cite

Wei, J., Zhu, Z., Niu, G., Liu, T., Liu, S., Sugiyama, M., & Liu, Y. (2026). Robust Learning from Noisily Labeled Long-Tailed Data via Fairness Regularizer. Proceedings of the AAAI Conference on Artificial Intelligence, 40(42), 35847-35856. https://doi.org/10.1609/aaai.v40i42.40899

Issue

Section

AAAI Technical Track on Philosophy and Ethics of AI