A Deep Architecture for Semantic Matching with Multiple Positional Sentence Representations

Authors

  • Shengxian Wan Chinese Academy of Sciences
  • Yanyan Lan Chinese Academy of Sciences
  • Jiafeng Guo Chinese Academy of Sciences
  • Jun Xu Chinese Academy of Sciences
  • Liang Pang Chinese Academy of Sciences
  • Xueqi Cheng Chinese Academy of Sciences

DOI:

https://doi.org/10.1609/aaai.v30i1.10342

Abstract

Matching natural language sentences is central for many applications such as information retrieval and question answering. Existing deep models rely on a single sentence representation or multiple granularity representations for matching. However, such methods cannot well capture the contextualized local information in the matching process. To tackle this problem, we present a new deep architecture to match two sentences with multiple positional sentence representations. Specifically, each positional sentence representation is a sentence representation at this position, generated by a bidirectional long short term memory (Bi-LSTM). The matching score is finally produced by aggregating interactions between these different positional sentence representations, through k-Max pooling and a multi-layer perceptron. Our model has several advantages: (1) By using Bi-LSTM, rich context of the whole sentence is leveraged to capture the contextualized local information in each positional sentence representation; (2) By matching with multiple positional sentence representations, it is flexible to aggregate different important contextualized local information in a sentence to support the matching; (3) Experiments on different tasks such as question answering and sentence completion demonstrate the superiority of our model.

Downloads

Published

2016-03-05

How to Cite

Wan, S., Lan, Y., Guo, J., Xu, J., Pang, L., & Cheng, X. (2016). A Deep Architecture for Semantic Matching with Multiple Positional Sentence Representations. Proceedings of the AAAI Conference on Artificial Intelligence, 30(1). https://doi.org/10.1609/aaai.v30i1.10342

Issue

Section

Technical Papers: NLP and Machine Learning