Multi-View Empowered Structural Graph Wordification for Language Models

Authors

  • Zipeng Liu College of Management and Economics, Tianjin University Laboratory of Computation and Analytics of Complex Management Systems (CACMS), Tianjin University
  • Likang Wu College of Management and Economics, Tianjin University Laboratory of Computation and Analytics of Complex Management Systems (CACMS), Tianjin University
  • Ming He AI Lab, Lenovo Research
  • Zhong Guan College of Management and Economics, Tianjin University Laboratory of Computation and Analytics of Complex Management Systems (CACMS), Tianjin University
  • Hongke Zhao College of Management and Economics, Tianjin University Laboratory of Computation and Analytics of Complex Management Systems (CACMS), Tianjin University
  • Nan Feng College of Management and Economics, Tianjin University Laboratory of Computation and Analytics of Complex Management Systems (CACMS), Tianjin University

DOI:

https://doi.org/10.1609/aaai.v39i23.34652

Abstract

Significant efforts have been dedicated to integrating the powerful Large Language Models (LLMs) with diverse modalities, particularly focusing on the fusion of language, vision and audio data. However, the graph-structured data, which is inherently rich in structural and domain-specific knowledge, has not yet been gracefully adapted to LLMs. Existing methods either describe the graph with raw text, suffering the loss of graph structural information, or feed Graph Neural Network (GNN) embeddings into LLMs at the cost of losing explainable prompt semantics. To bridge this gap, we introduce an end-to-end modality-aligning framework for LLM-graph alignment: Dual-Residual Vector Quantized-Variational AutoEncoder, namely Dr.E. Our approach is purposefully designed to facilitate token-level alignment with LLMs, enabling an effective translation of the intrinsic `language' of graphs into comprehensible natural language. We also manage to enhance LLMs' more robust structural understanding of graphs by incorporating multiple views of the central nodes based on their surrounding nodes at various distances. Our experimental evaluations on standard graph tasks demonstrate competitive performance against other state-of-the-art (SOTA) approaches. Additionally, our framework ensures certain visual interpretability, efficiency, and robustness, marking the promising successful endeavor to achieve token-level alignment between LLMs and GNNs.

Published

2025-04-11

How to Cite

Liu, Z., Wu, L., He, M., Guan, Z., Zhao, H., & Feng, N. (2025). Multi-View Empowered Structural Graph Wordification for Language Models. Proceedings of the AAAI Conference on Artificial Intelligence, 39(23), 24714–24722. https://doi.org/10.1609/aaai.v39i23.34652

Issue

Section

AAAI Technical Track on Natural Language Processing II