Audio-Oriented Multimodal Machine Comprehension via Dynamic Inter- and Intra-modality Attention

Authors

  • Zhiqi Huang Peking University
  • Fenglin Liu Peking University
  • Xian Wu Tencent AI Lab
  • Shen Ge Tencent AI Lab
  • Helin Wang Peking University
  • Wei Fan Tencent AI Lab
  • Yuexian Zou Peking University Peng Cheng Laboratory, China

Keywords:

Question Answering, Language Grounding & Multi-modal NLP, Applications

Abstract

While Machine Comprehension (MC) has attracted extensive research interests in recent years, existing approaches mainly belong to the category of Machine Reading Comprehension task which mines textual inputs (paragraphs and questions) to predict the answers (choices or text spans). However, there are a lot of MC tasks that accept audio input in addition to the textual input, e.g. English listening comprehension test. In this paper, we target the problem of Audio-Oriented Multimodal Machine Comprehension, and its goal is to answer questions based on the given audio and textual information. To solve this problem, we propose a Dynamic Inter- and Intra-modality Attention (DIIA) model to effectively fuse the two modalities (audio and textual). DIIA can work as an independent component and thus be easily integrated into existing MC models. Moreover, we further develop a Multimodal Knowledge Distillation (MKD) module to enable our multimodal MC model to accurately predict the answers based only on either the text or the audio. As a result, the proposed approach can handle various tasks including: Audio-Oriented Multimodal Machine Comprehension, Machine Reading Comprehension and Machine Listening Comprehension, in a single model, making fair comparisons possible between our model and the existing unimodal MC models. Experimental results and analysis prove the effectiveness of the proposed approaches. First, the proposed DIIA boosts the baseline models by up to 21.08% in terms of accuracy; Second, under the unimodal scenarios, the MKD module allows our multimodal MC model to significantly outperform the unimodal models by up to 18.87%, which are trained and tested with only audio or textual data.

Downloads

Published

2021-05-18

How to Cite

Huang, Z., Liu, F., Wu, X., Ge, S., Wang, H., Fan, W., & Zou, Y. (2021). Audio-Oriented Multimodal Machine Comprehension via Dynamic Inter- and Intra-modality Attention. Proceedings of the AAAI Conference on Artificial Intelligence, 35(14), 13098-13106. Retrieved from https://ojs.aaai.org/index.php/AAAI/article/view/17548

Issue

Section

AAAI Technical Track on Speech and Natural Language Processing I