Regularization Penalty Optimization for Addressing Data Quality Variance in OoD Algorithms

Authors

  • Runpeng Yu Tsinghua University
  • Hong Zhu Huawei Noah's Ark Lab
  • Kaican Li Huawei Noah's Ark Lab
  • Lanqing Hong Huawei Noah's Ark Lab
  • Rui Zhang ruizhang.info
  • Nanyang Ye Shanghai Jiao Tong University
  • Shao-Lun Huang TBSI
  • Xiuqiang He Huawei Noah's Ark Lab

DOI:

https://doi.org/10.1609/aaai.v36i8.20877

Keywords:

Machine Learning (ML)

Abstract

Due to the poor generalization performance of traditional empirical risk minimization (ERM) in the case of distributional shift, Out-of-Distribution (OoD) generalization algorithms receive increasing attention. However, OoD generalization algorithms overlook the great variance in the quality of training data, which significantly compromises the accuracy of these methods. In this paper, we theoretically reveal the relationship between training data quality and algorithm performance, and analyze the optimal regularization scheme for Lipschitz regularized invariant risk minimization. A novel algorithm is proposed based on the theoretical results to alleviate the influence of low quality data at both the sample level and the domain level. The experiments on both the regression and classification benchmarks validate the effectiveness of our method with statistical significance.

Downloads

Published

2022-06-28

How to Cite

Yu, R., Zhu, H., Li, K., Hong, L., Zhang, R., Ye, N., Huang, S.-L., & He, X. (2022). Regularization Penalty Optimization for Addressing Data Quality Variance in OoD Algorithms. Proceedings of the AAAI Conference on Artificial Intelligence, 36(8), 8945-8953. https://doi.org/10.1609/aaai.v36i8.20877

Issue

Section

AAAI Technical Track on Machine Learning III