GradPU: Positive-Unlabeled Learning via Gradient Penalty and Positive Upweighting
DOI:
https://doi.org/10.1609/aaai.v37i6.25889Keywords:
ML: Classification and Regression, CV: Object Detection & Categorization, ML: Semi-Supervised Learning, ML: Unsupervised & Self-Supervised LearningAbstract
Positive-unlabeled learning is an essential problem in many real-world applications with only labeled positive and unlabeled data, especially when the negative samples are difficult to identify. Most existing positive-unlabeled learning methods will inevitably overfit the positive class to some extent due to the existence of unidentified positive samples. This paper first analyzes the overfitting problem and proposes to bound the generalization errors via Wasserstein distances. Based on that, we develop a simple yet effective positive-unlabeled learning method, GradPU, which consists of two key ingredients: A gradient-based regularizer that penalizes the gradient norms in the interpolated data region, which improves the generalization of positive class; An unnormalized upweighting mechanism that assigns larger weights to those positive samples that are hard, not-well-fitted and less frequently labeled. It enforces the training error of each positive sample to be small and increases the robustness to the labeling bias. We evaluate our proposed GradPU on three datasets: MNIST, FashionMNIST, and CIFAR10. The results demonstrate that GradPU achieves state-of-the-art performance on both unbiased and biased positive labeling scenarios.Downloads
Published
2023-06-26
How to Cite
Dai, S., Li, X., Zhou, Y., Ye, X., & Liu, T. (2023). GradPU: Positive-Unlabeled Learning via Gradient Penalty and Positive Upweighting. Proceedings of the AAAI Conference on Artificial Intelligence, 37(6), 7296-7303. https://doi.org/10.1609/aaai.v37i6.25889
Issue
Section
AAAI Technical Track on Machine Learning I