Adversarial Weight Perturbation Improves Generalization in Graph Neural Networks

Authors

  • Yihan Wu University of Pittsburgh
  • Aleksandar Bojchevski CISPA Helmholtz Center for Information Security
  • Heng Huang University of Pittsburgh

DOI:

https://doi.org/10.1609/aaai.v37i9.26239

Keywords:

ML: Graph-based Machine Learning, ML: Adversarial Learning & Robustness

Abstract

A lot of theoretical and empirical evidence shows that the flatter local minima tend to improve generalization. Adversarial Weight Perturbation (AWP) is an emerging technique to efficiently and effectively find such minima. In AMP we minimize the loss w.r.t. a bounded worst-case perturbation of the model parameters thereby favoring local minima with a small loss in a neighborhood around them. The benefits of AWP, and more generally the connections between flatness and generalization, have been extensively studied for i.i.d. data such as images. In this paper, we extensively study this phenomenon for graph data. Along the way, we first derive a generalization bound for non-i.i.d. node classification tasks. Then we identify a vanishing-gradient issue with all existing formulations of AWP and we propose a new Weighted Truncated AWP (WT-AWP) to alleviate this issue. We show that regularizing graph neural networks with WT-AWP consistently improves both natural and robust generalization across many different graph learning tasks and models.

Downloads

Published

2023-06-26

How to Cite

Wu, Y., Bojchevski, A., & Huang, H. (2023). Adversarial Weight Perturbation Improves Generalization in Graph Neural Networks. Proceedings of the AAAI Conference on Artificial Intelligence, 37(9), 10417-10425. https://doi.org/10.1609/aaai.v37i9.26239

Issue

Section

AAAI Technical Track on Machine Learning IV