SPRING: Situated Conversation Agent Pretrained with Multimodal Questions from Incremental Layout Graph
DOI:
https://doi.org/10.1609/aaai.v37i11.26562Keywords:
SNLP: Conversational AI/Dialogue Systems, CV: Multi-modal Vision, SNLP: Generation, SNLP: Language Models, SNLP: Question AnsweringAbstract
Existing multimodal conversation agents have shown impressive abilities to locate absolute positions or retrieve attributes in simple scenarios, but they fail to perform well when complex relative positions and information alignments are involved, which poses a bottleneck in response quality. In this paper, we propose a Situated Conversation Agent Pretrained with Multimodal Questions from Incremental Layout Graph (SPRING) with abilities of reasoning multi-hops spatial relations and connecting them with visual attributes in crowded situated scenarios. Specifically, we design two types of Multimodal Question Answering (MQA) tasks to pretrain the agent. All QA pairs utilized during pretraining are generated from novel Increment Layout Graphs (ILG). QA pair difficulty labels automatically annotated by ILG are used to promote MQA-based Curriculum Learning. Experimental results verify the SPRING's effectiveness, showing that it significantly outperforms state-of-the-art approaches on both SIMMC 1.0 and SIMMC 2.0 datasets. We release our code and data at https://github.com/LYX0501/SPRING.Downloads
Published
2023-06-26
How to Cite
Long, Y., Hui, B., Ye, F., Li, Y., Han, Z., Yuan, C., Li, Y., & Wang, X. (2023). SPRING: Situated Conversation Agent Pretrained with Multimodal Questions from Incremental Layout Graph. Proceedings of the AAAI Conference on Artificial Intelligence, 37(11), 13309-13317. https://doi.org/10.1609/aaai.v37i11.26562
Issue
Section
AAAI Technical Track on Speech & Natural Language Processing