LLMTM: Benchmarking and Optimizing LLMs for Temporal Motif Analysis in Dynamic Graphs

Authors

  • Bing Hao School of New Media and Communication, Tianjin University, China School of Computer Science and Technology, Beihang University, China
  • Minglai Shao School of New Media and Communication, Tianjin University, China Key Lab of Education Blockchain and Intelligent Technology, Ministry of Education, Guangxi Normal University, China
  • Zengyi Wo School of New Media and Communication, Tianjin University, China
  • Yunlong Chu School of New Media and Communication, Tianjin University, China
  • Yuhang Liu School of New Media and Communication, Tianjin University, China
  • Ruijie Wang School of Computer Science and Technology, Beihang University, China

DOI:

https://doi.org/10.1609/aaai.v40i17.38500

Abstract

The widespread application of Large Language Models (LLMs) has motivated a growing interest in their capacity for processing dynamic graphs. Temporal motifs, as an elementary unit and important local property of dynamic graphs which can directly reflect anomalies and unique phenomena, are essential for understanding their evolutionary dynamics and structural features. However, leveraging LLMs for temporal motif analysis on dynamic graphs remains relatively unexplored. In this paper, we systematically study LLM performance on temporal motif-related tasks. Specifically, we propose a comprehensive benchmark, LLMTM (Large Language Models in Temporal Motifs), which includes six tailored tasks across nine temporal motif types. We then conduct extensive experiments to analyze the impacts of different prompting techniques and LLMs (including nine models: openPangu-7B, the DeepSeek-R1-Distill-Qwen series, Qwen2.5-32B-Instruct, GPT-4o-mini, DeepSeek-R1, and o3) on model performance. Informed by our benchmark findings, we develop a tool-augmented LLM agent that leverages precisely engineered prompts to solve these tasks with high accuracy. Nevertheless, the high accuracy of the agent incurs a substantial cost. To address this trade-off, we propose a simple yet effective structure-aware dispatcher that considers both the dynamic graph's structural properties and the LLM's cognitive load to intelligently dispatch queries between the standard LLM prompting and the more powerful agent. Our experiments demonstrate that the structure-aware dispatcher effectively maintains high accuracy while reducing cost.

Published

2026-03-14

How to Cite

Hao, B., Shao, M., Wo, Z., Chu, Y., Liu, Y., & Wang, R. (2026). LLMTM: Benchmarking and Optimizing LLMs for Temporal Motif Analysis in Dynamic Graphs. Proceedings of the AAAI Conference on Artificial Intelligence, 40(17), 14802–14810. https://doi.org/10.1609/aaai.v40i17.38500

Issue

Section

AAAI Technical Track on Data Mining & Knowledge Management I