Learning Structured Embeddings of Knowledge Bases

Authors

  • Antoine Bordes CNRS, Université de Technologie de Compiègne
  • Jason Weston Google, Inc.
  • Ronan Collobert IDIAP
  • Yoshua Bengio Université de Montréal

Abstract

Many Knowledge Bases (KBs) are now readily available and encompass colossal quantities of information thanks to either a long-term funding effort (e.g. WordNet, OpenCyc) or a collaborative process (e.g. Freebase, DBpedia). However, each of them is based on a different rigorous symbolic framework which makes it hard to use their data in other systems. It is unfortunate because such rich structured knowledge might lead to a huge leap forward in many other areas of AI like nat- ural language processing (word-sense disambiguation, natural language understanding, ...), vision (scene classification, image semantic annotation, ...) or collaborative filtering. In this paper, we present a learning process based on an innovative neural network architecture designed to embed any of these symbolic representations into a more flexible continuous vector space in which the original knowledge is kept and enhanced. These learnt embeddings would allow data from any KB to be easily used in recent machine learning meth- ods for prediction and information retrieval. We illustrate our method on WordNet and Freebase and also present a way to adapt it to knowledge extraction from raw text.

Downloads

Published

2011-08-04

How to Cite

Bordes, A., Weston, J., Collobert, R., & Bengio, Y. (2011). Learning Structured Embeddings of Knowledge Bases. Proceedings of the AAAI Conference on Artificial Intelligence, 25(1), 301-306. Retrieved from https://ojs.aaai.org/index.php/AAAI/article/view/7917

Issue

Section

AAAI Technical Track: Machine Learning