LGMRec: Local and Global Graph Learning for Multimodal Recommendation

Authors

  • Zhiqiang Guo School of Computer Science and Technology, Huazhong University of Science and Technology
  • Jianjun Li School of Computer Science and Technology, Huazhong University of Science and Technology
  • Guohui Li School of Software Engineering, Huazhong University of Science and Technology
  • Chaoyang Wang Wuhan Digital Engineering Institute
  • Si Shi Guangdong Laboratory of Artificial Intelligence and Digital Economy(SZ)
  • Bin Ruan School of Computer Science and Technology, Huazhong University of Science and Technology

DOI:

https://doi.org/10.1609/aaai.v38i8.28688

Keywords:

DMKM: Recommender Systems, DMKM: Mining of Visual, Multimedia & Multimodal Data

Abstract

The multimodal recommendation has gradually become the infrastructure of online media platforms, enabling them to provide personalized service to users through a joint modeling of user historical behaviors (e.g., purchases, clicks) and item various modalities (e.g., visual and textual). The majority of existing studies typically focus on utilizing modal features or modal-related graph structure to learn user local interests. Nevertheless, these approaches encounter two limitations: (1) Shared updates of user ID embeddings result in the consequential coupling between collaboration and multimodal signals; (2) Lack of exploration into robust global user interests to alleviate the sparse interaction problems faced by local interest modeling. To address these issues, we propose a novel Local and Global Graph Learning-guided Multimodal Recommender (LGMRec), which jointly models local and global user interests. Specifically, we present a local graph embedding module to independently learn collaborative-related and modality-related embeddings of users and items with local topological relations. Moreover, a global hypergraph embedding module is designed to capture global user and item embeddings by modeling insightful global dependency relations. The global embeddings acquired within the hypergraph embedding space can then be combined with two decoupled local embeddings to improve the accuracy and robustness of recommendations. Extensive experiments conducted on three benchmark datasets demonstrate the superiority of our LGMRec over various state-of-the-art recommendation baselines, showcasing its effectiveness in modeling both local and global user interests.

Published

2024-03-24

How to Cite

Guo, Z., Li, J., Li, G., Wang, C., Shi, S., & Ruan, B. (2024). LGMRec: Local and Global Graph Learning for Multimodal Recommendation. Proceedings of the AAAI Conference on Artificial Intelligence, 38(8), 8454-8462. https://doi.org/10.1609/aaai.v38i8.28688

Issue

Section

AAAI Technical Track on Data Mining & Knowledge Management