Multi-Dimension Attention for Multi-Turn Dialog Generation (Student Abstract)
Keywords:Sequence Transduction, Dialog Generation, Neural Machine, Attention Mechanism, Recosa, Hred
AbstractWe present a generative neural model for open and multi-turn dialog response generation that relies on a multi-dimension attention process to account for the semantic interdependence between the generated words and the conversational history, so as to identify all the words and utterances that influence each generated response. The performance of the model is evaluated on the wide scope DailyDialog corpus and a comparison is made with two other generative neural architectures, using several machine metrics. The results show that the proposed model improves the state of the art for generation accuracy, and its multi-dimension attention allows for a more detailed tracking of the influential words and utterances in the dialog history for response explainability by the dialog history.
How to Cite
Belainine, B., Sadat, F., & Boukadoum, M. (2022). Multi-Dimension Attention for Multi-Turn Dialog Generation (Student Abstract). Proceedings of the AAAI Conference on Artificial Intelligence, 36(11), 12909-12910. https://doi.org/10.1609/aaai.v36i11.21591
AAAI Student Abstract and Poster Program