Neural Knowledge Acquisition via Mutual Attention Between Knowledge Graph and Text

Authors

  • Xu Han Tsinghua University
  • Zhiyuan Liu Tsinghua University
  • Maosong Sun Tsinghua University

DOI:

https://doi.org/10.1609/aaai.v32i1.11927

Keywords:

Joint Learning, Mutual Attention, Knowledge Graph Completion, Relation Extraction

Abstract

We propose a general joint representation learning framework for knowledge acquisition (KA) on two tasks, knowledge graph completion (KGC) and relation extraction (RE) from text. In this framework, we learn representations of knowledge graphs (KGs) and text within a unified parameter sharing semantic space. To achieve better fusion, we propose an effective mutual attention between KGs and text. The reciprocal attention mechanism enables us to highlight important features and perform better KGC and RE. Different from conventional joint models, no complicated linguistic analysis or strict alignments between KGs and text are required to train our models. Experiments on relation extraction and entity link prediction show that models trained under our joint framework are significantly improved in comparison with other baselines. Most existing methods for KGC and RE can be easily integrated into our framework due to its flexible architectures. The source code of this paper can be obtained from https://github.com/thunlp/JointNRE.

Downloads

Published

2018-04-26

How to Cite

Han, X., Liu, Z., & Sun, M. (2018). Neural Knowledge Acquisition via Mutual Attention Between Knowledge Graph and Text. Proceedings of the AAAI Conference on Artificial Intelligence, 32(1). https://doi.org/10.1609/aaai.v32i1.11927

Issue

Section

Main Track: NLP and Knowledge Representation