Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models

Authors

  • Iulian Serban University of Montreal
  • Alessandro Sordoni University of Montreal
  • Yoshua Bengio University of Montreal
  • Aaron Courville University of Montreal
  • Joelle Pineau McGill University

DOI:

https://doi.org/10.1609/aaai.v30i1.9883

Keywords:

Dialogue Systems, Cognitive Systems, Neural Networks, Generative Probabilistic Models, Word Embeddings, Transfer Learning

Abstract

We investigate the task of building open domain, conversational dialogue systems based on large dialogue corpora using generative models. Generative models produce system responses that are autonomously generated word-by-word, opening up the possibility for realistic, flexible interactions. In support of this goal, we extend the recently proposed hierarchical recurrent encoder-decoder neural network to the dialogue domain, and demonstrate that this model is competitive with state-of-the-art neural language models and back-off n-gram models. We investigate the limitations of this and similar approaches, and show how its performance can be improved by bootstrapping the learning from a larger question-answer pair corpus and from pretrained word embeddings.

Downloads

Published

2016-03-05

How to Cite

Serban, I., Sordoni, A., Bengio, Y., Courville, A., & Pineau, J. (2016). Building End-To-End Dialogue Systems Using Generative Hierarchical Neural Network Models. Proceedings of the AAAI Conference on Artificial Intelligence, 30(1). https://doi.org/10.1609/aaai.v30i1.9883