Deep Hypergraph Neural Networks with Tight Framelets

Authors

  • Ming Li Zhejiang Key Laboratory of Intelligent Education Technology and Application, Zhejiang Normal University Zhejiang Institute of Optoelectronics
  • Yujie Fang School of Computer Science and Technology, Zhejiang Normal University Zhejiang Key Laboratory of Intelligent Education Technology and Application, Zhejiang Normal University
  • Yi Wang School of Computer Science and Technology, Zhejiang Normal University Zhejiang Key Laboratory of Intelligent Education Technology and Application, Zhejiang Normal University
  • Han Feng Department of Mathematics, City University of Hong Kong
  • Yongchun Gu Zhejiang Key Laboratory of Intelligent Education Technology and Application, Zhejiang Normal University
  • Lu Bai School of Artificial Intelligence, and Engineering Research Center of Intelligent Technology and Educational Application, Ministry of Education, Beijing Normal University
  • Pietro Liò Department of Computer Science and Technology, Cambridge University

DOI:

https://doi.org/10.1609/aaai.v39i17.34023

Abstract

Hypergraphs provide a flexible framework for modeling high-order (complex) interactions among multiple entities, extending beyond traditional pairwise correlations in graph structures. However, deep hypergraph neural networks (HGNNs) often face the challenge of oversmoothing with increasing depth, similar to issues in graph neural networks (GNNs). While oversmoothing in GNNs has been extensively studied, its implications in relation to hypergraphs are less explored. This paper addresses this gap by first theoretically exploring the reasons behind oversmoothing in deep HGNNs. Our novel insights suggest that a spectral-based hypergraph convolution, equipped with both low-pass and high-pass filters, can potentially mitigate these effects. Motivated by these findings, we introduce FrameHGNN, a framework that utilizes framelet-based hypergraph convolutions integrating tight framelet transforms with both low-pass and high-pass components, as well as the commonly used strategies in designing deep GNN architecture: initial residual and identity mappings. The experiment results on diverse benchmark datasets demonstrate that FrameHGNN outperforms several state-of-the-art models, effectively reducing oversmoothing while improving predictive accuracy. Our contributions not only advance the theoretical understanding of deep hypergraph learning but also provide a practical spectral-based approach for HGNNs, emphasizing the design of multifrequency channels.

Published

2025-04-11

How to Cite

Li, M., Fang, Y., Wang, Y., Feng, H., Gu, Y., Bai, L., & Liò, P. (2025). Deep Hypergraph Neural Networks with Tight Framelets. Proceedings of the AAAI Conference on Artificial Intelligence, 39(17), 18385-18392. https://doi.org/10.1609/aaai.v39i17.34023

Issue

Section

AAAI Technical Track on Machine Learning III