Kernelized Online Imbalanced Learning with Fixed Budgets

Authors

  • Junjie Hu The Chinese University of Hong Kong
  • Haiqin Yang The Chinese University of Hong Kong
  • Irwin King The Chinese University of Hong Kong
  • Michael Lyu The Chinese University of Hong Kong
  • Anthony Man-Cho So The Chinese University of Hong Kong

Abstract

Online learning from imbalanced streaming data to capture the nonlinearity and heterogeneity of the data is significant in machine learning and data mining. To tackle this problem, we propose a kernelized online imbalanced learning (KOIL) algorithm to directly maximize the area under the ROC curve (AUC). We address two more challenges: 1) How to control the number of support vectors without sacrificing model performance; and 2) how to restrict the fluctuation of the learned decision function to attain smooth updating. To this end, we introduce two buffers with fixed budgets (buffer sizes) for positive class and negative class, respectively, to store the learned support vectors, which can allow us to capture the global information of the decision boundary. When determining the weight of a new support vector, we confine its influence only to its $k$-nearest opposite support vectors. This can restrict the effect of new instances and prevent the harm of outliers. More importantly, we design a sophisticated scheme to compensate the model after replacement is conducted when either buffer is full. With this compensation, the learned model approaches the one learned with infinite budgets. We present both theoretical analysis and extensive experimental comparison to demonstrate the effectiveness of our proposed KOIL.

Downloads

Published

2015-02-21

How to Cite

Hu, J., Yang, H., King, I., Lyu, M., & So, A. M.-C. (2015). Kernelized Online Imbalanced Learning with Fixed Budgets. Proceedings of the AAAI Conference on Artificial Intelligence, 29(1). Retrieved from https://ojs.aaai.org/index.php/AAAI/article/view/9587

Issue

Section

Main Track: Novel Machine Learning Algorithms