Subgraph Invariant Learning Towards Large-Scale Graph Node Classification

Authors

  • Leilei Wang Guangdong Laboratory of Artificial Intelligence and Digital Economy (SZ), Shenzhen, China College of Computer Science and Software Engineering, Shenzhen University, China
  • Si Shi Guangdong Laboratory of Artificial Intelligence and Digital Economy (SZ), Shenzhen, China
  • Fei Ma Guangdong Laboratory of Artificial Intelligence and Digital Economy (SZ), Shenzhen, China
  • Fei Richard Yu College of Computer Science and Software Engineering, Shenzhen University, China School of Information Technology, Carleton University, Canada
  • Pengteng Li HKUST (GZ), AI Thrust
  • Ying Tiffany He College of Computer Science and Software Engineering, Shenzhen University, China

DOI:

https://doi.org/10.1609/aaai.v39i20.35412

Abstract

Graph Neural Networks (GNNs) have shown efficacy in graph node classification, but face computational challenges on large-scale graphs. Although existing graph reduction methods address these issues, they still require high computational resources and fail to prioritize robust performance on out-of-distribution data. To tackle these challenges, we introduce the subgraph invariant learning paradigm, inspired by the small-world phenomenon. This approach enables models trained on specific subgraphs to generalize across diverse subgraphs, reducing computational demands, and enhancing scalability. To promote generalization, we maximize the invariance log-likelihood by deriving a theoretical lower bound of it and formulating the InVar loss. This loss minimizes the discrepancy between node representations and their corresponding invariance representations while maximizing the entropy of the node representation. In response to InVar loss, we propose the Invariance Facilitation Model (IFM), comprising the Invariance Representation Encoder (IRE) and Node Representation Encoder (NRE). IRE, capturing the invariance representations, utilizes Invariance ATTention (InvarATT) to compress long-range dependencies, while NRE learns the node representation, by integrating invariance representations via Telematic ATTention (TeleATT) and exchanging local information within each subgraph through GNNs. Evaluations on four large-scale graph datasets demonstrate the effectiveness, computational efficiency, and interpretability of IFM for large-scale graph node classification.

Downloads

Published

2025-04-11

How to Cite

Wang, L., Shi, S., Ma, F., Yu, F. R., Li, P., & He, Y. T. (2025). Subgraph Invariant Learning Towards Large-Scale Graph Node Classification. Proceedings of the AAAI Conference on Artificial Intelligence, 39(20), 21144–21152. https://doi.org/10.1609/aaai.v39i20.35412

Issue

Section

AAAI Technical Track on Machine Learning VI