Imbalanced Label Distribution Learning
DOI:
https://doi.org/10.1609/aaai.v37i9.26341Keywords:
ML: Multi-Class/Multi-Label Learning & Extreme ClassificationAbstract
Label distribution covers a certain number of labels, representing the degree to which each label describes an instance. The learning process on the instances labeled by label distributions is called Label Distribution Learning (LDL). Although LDL has been applied successfully to many practical applications, one problem with existing LDL methods is that they are limited to data with balanced label information. However, annotation information in real-world data often exhibits imbalanced distributions, which significantly degrades the performance of existing methods. In this paper, we investigate the Imbalanced Label Distribution Learning (ILDL) problem. To handle this challenging problem, we delve into the characteristics of ILDL and empirically find that the representation distribution shift is the underlying reason for the performance degradation of existing methods. Inspired by this finding, we present a novel method named Representation Distribution Alignment (RDA). RDA aligns the distributions of feature representations and label representations to alleviate the impact of the distribution gap between the training set and the test set caused by the imbalance issue. Extensive experiments verify the superior performance of RDA. Our work fills the gap in benchmarks and techniques for practical ILDL problems.Downloads
Published
2023-06-26
How to Cite
Zhao, X., An, Y., Xu, N., Wang, J., & Geng, X. (2023). Imbalanced Label Distribution Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 37(9), 11336-11344. https://doi.org/10.1609/aaai.v37i9.26341
Issue
Section
AAAI Technical Track on Machine Learning IV