Scalable Spatiotemporal Graph Neural Networks

Authors

  • Andrea Cini The Swiss AI Lab IDSIA, Università della Svizzera italiana
  • Ivan Marisca The Swiss AI Lab IDSIA, Università della Svizzera italiana
  • Filippo Maria Bianchi UiT the Arctic University of Norway NORCE Norwegian Research Centre
  • Cesare Alippi The Swiss AI Lab IDSIA, Università della Svizzera italiana Politecnico di Milano

DOI:

https://doi.org/10.1609/aaai.v37i6.25880

Keywords:

ML: Graph-based Machine Learning, ML: Time-Series/Data Streams, DMKM: Mining of Spatial, Temporal or Spatio-Temporal Data, ML: Scalability of ML Systems, ML: Deep Neural Architectures

Abstract

Neural forecasting of spatiotemporal time series drives both research and industrial innovation in several relevant application domains. Graph neural networks (GNNs) are often the core component of the forecasting architecture. However, in most spatiotemporal GNNs, the computational complexity scales up to a quadratic factor with the length of the sequence times the number of links in the graph, hence hindering the application of these models to large graphs and long temporal sequences. While methods to improve scalability have been proposed in the context of static graphs, few research efforts have been devoted to the spatiotemporal case. To fill this gap, we propose a scalable architecture that exploits an efficient encoding of both temporal and spatial dynamics. In particular, we use a randomized recurrent neural network to embed the history of the input time series into high-dimensional state representations encompassing multi-scale temporal dynamics. Such representations are then propagated along the spatial dimension using different powers of the graph adjacency matrix to generate node embeddings characterized by a rich pool of spatiotemporal features. The resulting node embeddings can be efficiently pre-computed in an unsupervised manner, before being fed to a feed-forward decoder that learns to map the multi-scale spatiotemporal representations to predictions. The training procedure can then be parallelized node-wise by sampling the node embeddings without breaking any dependency, thus enabling scalability to large networks. Empirical results on relevant datasets show that our approach achieves results competitive with the state of the art, while dramatically reducing the computational burden.

Downloads

Published

2023-06-26

How to Cite

Cini, A., Marisca, I., Bianchi, F. M., & Alippi, C. (2023). Scalable Spatiotemporal Graph Neural Networks. Proceedings of the AAAI Conference on Artificial Intelligence, 37(6), 7218-7226. https://doi.org/10.1609/aaai.v37i6.25880

Issue

Section

AAAI Technical Track on Machine Learning I