Graph2Video: Leveraging Video Models to Model Dynamic Graph Evolution
DOI:
https://doi.org/10.1609/aaai.v40i18.38557Abstract
Dynamic graphs are common in real‑world systems such as social media, recommender systems, and traffic networks. Existing dynamic graph models for link prediction often fall short in capturing the full complexity of temporal evolution. They tend to overlook fine‑grained variations in interaction order, struggle with dependencies that span long time horizons, and provide limited modeling of pair‑specific relational dynamics. To address those challenges, we propose Graph2Video, a video‑inspired framework that views the temporal neighborhood of a target link as a sequence of “graph frames”. By stacking temporally ordered subgraph frames into a “graph video”, Graph2Video leverages the inductive biases of video foundation models to capture both fine-grained local variations and long-range temporal dynamics. It generates a link-level embedding that serves as a lightweight, plug-and-play, link-centric memory unit. This embedding integrates seamlessly into existing dynamic graph encoders, effectively addressing the limitations of prior approaches. Extensive experiments on benchmark datasets show that Graph2Video outperforms state‑of‑the‑art baselines in the link prediction task on most cases. The results highlight that borrowing spatio‑temporal modeling techniques from computer vision provides a principled and effective avenue for advancing dynamic graph learning.Downloads
Published
2026-03-14
How to Cite
Liu, H., Wei, Y., Xing, F., Derr, T., Han, H., & Zhang, Y. (2026). Graph2Video: Leveraging Video Models to Model Dynamic Graph Evolution. Proceedings of the AAAI Conference on Artificial Intelligence, 40(18), 15315–15323. https://doi.org/10.1609/aaai.v40i18.38557
Issue
Section
AAAI Technical Track on Data Mining & Knowledge Management II