Utility-Oriented Knowledge Graph Accuracy Estimation with Limited Annotations: A Case Study on DBpedia

Authors

  • Stefano Marchesin Department of Information Engineering, University of Padua, Padua, Italy
  • Gianmaria Silvello Department of Information Engineering, University of Padua, Padua, Italy
  • Omar Alonso Amazon, Palo Alto, California, USA

DOI:

https://doi.org/10.1609/hcomp.v12i1.31605

Abstract

Knowledge Graphs (KGs) are essential for applications like search, recommendation, and virtual assistants, where their accuracy directly impacts effectiveness. However, due to their large-scale and ever-evolving nature, it is impractical to manually evaluate all KG contents. We propose a framework that employs sampling, estimation, and active learning to audit KG accuracy in a cost-effective manner. The framework prioritizes KG facts based on their utility to downstream tasks. We applied the framework to DBpedia and gathered annotations from both expert and layman annotators. We also explored the potential of Large Language Models (LLMs) as KG evaluators, showing that while they can perform comparably to low-quality human annotators, they tend to overestimate KG accuracy. As such, LLMs are currently insufficient to replace human crowdworkers in the evaluation process. The results also provide insights into the scalability of methods for auditing KGs.

Downloads

Published

2024-10-14

How to Cite

Marchesin, S., Silvello, G., & Alonso, O. (2024). Utility-Oriented Knowledge Graph Accuracy Estimation with Limited Annotations: A Case Study on DBpedia. Proceedings of the AAAI Conference on Human Computation and Crowdsourcing, 12(1), 105-114. https://doi.org/10.1609/hcomp.v12i1.31605