RQUERY: Rewriting Natural Language Queries on Knowledge Graphs to Alleviate the Vocabulary Mismatch Problem

Authors

  • Saeedeh Shekarpour Wright State University
  • Edgard Marx Universität Leipzig
  • Sören Auer University of Bonn
  • Amit Sheth Wright State University

DOI:

https://doi.org/10.1609/aaai.v31i1.11131

Keywords:

Query rewriting, Hidden Markov model, n-gram language model, triple-based co-occurence

Abstract

For non-expert users, a textual query is the most popular and simple means for communicating with a retrieval or question answering system.However, there is a risk of receiving queries which do not match with the background knowledge.Query expansion and query rewriting are solutions for this problem but they are in danger of potentially yielding a large number of irrelevant words, which in turn negatively influences runtime as well as accuracy.In this paper, we propose a new method for automatic rewriting input queries on graph-structured RDF knowledge bases.We employ a Hidden Markov Model to determine the most suitable derived words from linguistic resources.We introduce the concept of triple-based co-occurrence for recognizing co-occurred words in RDF data.This model was bootstrapped with three statistical distributions.Our experimental study demonstrates the superiority of the proposed approach to the traditional n-gram model.

Downloads

Published

2017-02-12

How to Cite

Shekarpour, S., Marx, E., Auer, S., & Sheth, A. (2017). RQUERY: Rewriting Natural Language Queries on Knowledge Graphs to Alleviate the Vocabulary Mismatch Problem. Proceedings of the AAAI Conference on Artificial Intelligence, 31(1). https://doi.org/10.1609/aaai.v31i1.11131

Issue

Section

Main Track: Search and Constraint Satisfaction