Assessing LLMs for Serendipity Discovery in Knowledge Graphs: A Case for Drug Repurposing

Authors

  • Mengying Wang Case Western Reserve University
  • Chenhui Ma Case Western Reserve University
  • Ao Jiao Case Western Reserve University
  • Tuo Liang Case Western Reserve University
  • Pengjun Lu Case Western Reserve University
  • Shrinidhi Hegde Case Western Reserve University
  • Yu Yin Case Western Reserve University
  • Evren Gurkan-Cavusoglu Case Western Reserve University
  • Yinghui Wu Case Western Reserve University

DOI:

https://doi.org/10.1609/aaai.v40i19.38618

Abstract

Large Language Models (LLMs) have greatly advanced knowledge graph question answering (KGQA), yet existing systems are typically optimized for returning highly relevant but predictable answers. A missing yet desired capacity is to exploit LLMs to suggest surprise and novel ("serendipitious") answers. In this paper, we formally define the serendipity-aware KGQA task and propose the SerenQA framework to evaluate LLMs' ability to uncover unexpected insights in scientific KGQA tasks. SerenQA includes a rigorous serendipity metric based on relevance, novelty, and surprise, along with an expert-annotated benchmark derived from the Clinical Knowledge Graph for drug repurposing. Additionally, it features a structured evaluation pipeline encompassing three subtasks: knowledge retrieval, subgraph reasoning, and serendipity exploration. Our experiments reveal that while state-of-the-art LLMs perform well on retrieval, they still struggle to identify genuinely surprising and valuable discoveries, underscoring a significant room for future research.

Published

2026-03-14

How to Cite

Wang, M., Ma, C., Jiao, A., Liang, T., Lu, P., Hegde, S., … Wu, Y. (2026). Assessing LLMs for Serendipity Discovery in Knowledge Graphs: A Case for Drug Repurposing. Proceedings of the AAAI Conference on Artificial Intelligence, 40(19), 15860–15867. https://doi.org/10.1609/aaai.v40i19.38618

Issue

Section

AAAI Technical Track on Data Mining & Knowledge Management III