Multi-Dimension Attention for Multi-Turn Dialog Generation (Student Abstract)

Authors

  • Billal Belainine University of Quebec in Montreal
  • Fatiha Sadat University of Quebec in Montreal
  • Mounir Boukadoum University of Quebec in Montreal

DOI:

https://doi.org/10.1609/aaai.v36i11.21591

Keywords:

Sequence Transduction, Dialog Generation, Neural Machine, Attention Mechanism, Recosa, Hred

Abstract

We present a generative neural model for open and multi-turn dialog response generation that relies on a multi-dimension attention process to account for the semantic interdependence between the generated words and the conversational history, so as to identify all the words and utterances that influence each generated response. The performance of the model is evaluated on the wide scope DailyDialog corpus and a comparison is made with two other generative neural architectures, using several machine metrics. The results show that the proposed model improves the state of the art for generation accuracy, and its multi-dimension attention allows for a more detailed tracking of the influential words and utterances in the dialog history for response explainability by the dialog history.

Downloads

Published

2022-06-28

How to Cite

Belainine, B., Sadat, F., & Boukadoum, M. (2022). Multi-Dimension Attention for Multi-Turn Dialog Generation (Student Abstract). Proceedings of the AAAI Conference on Artificial Intelligence, 36(11), 12909-12910. https://doi.org/10.1609/aaai.v36i11.21591