Semantic Structure-Based Word Embedding by Incorporating Concept Convergence and Word Divergence

Authors

  • Qian Liu Beijing Institute of Technology
  • Heyan Huang Beijing Institute of Technology
  • Guangquan Zhang University of Technology Sydney
  • Yang Gao Beijing Institute of Technology
  • Junyu Xuan University of Technology Sydney
  • Jie Lu University of Technology Sydney

DOI:

https://doi.org/10.1609/aaai.v32i1.11942

Keywords:

Word embedding, nature language processing

Abstract

Representing the semantics of words is a fundamental task in text processing. Several research studies have shown that text and knowledge bases (KBs) are complementary sources for word embedding learning. Most existing methods only consider relationships within word-pairs in the usage of KBs. We argue that the structural information of well-organized words within the KBs is able to convey more effective and stable knowledge in capturing semantics of words. In this paper, we propose a semantic structure-based word embedding method, and introduce concept convergence and word divergence to reveal semantic structures in the word embedding learning process. To assess the effectiveness of our method, we use WordNet for training and conduct extensive experiments on word similarity, word analogy, text classification and query expansion. The experimental results show that our method outperforms state-of-the-art methods, including the methods trained solely on the corpus, and others trained on the corpus and the KBs.

Downloads

Published

2018-04-27

How to Cite

Liu, Q., Huang, H., Zhang, G., Gao, Y., Xuan, J., & Lu, J. (2018). Semantic Structure-Based Word Embedding by Incorporating Concept Convergence and Word Divergence. Proceedings of the AAAI Conference on Artificial Intelligence, 32(1). https://doi.org/10.1609/aaai.v32i1.11942