A Natural Language Corpus of Common Grounding under Continuous and Partially-Observable Context

Authors

  • Takuma Udagawa University of Tokyo
  • Akiko Aizawa National Institute of Informatics

DOI:

https://doi.org/10.1609/aaai.v33i01.33017120

Abstract

Common grounding is the process of creating, repairing and updating mutual understandings, which is a critical aspect of sophisticated human communication. However, traditional dialogue systems have limited capability of establishing common ground, and we also lack task formulations which introduce natural difficulty in terms of common grounding while enabling easy evaluation and analysis of complex models. In this paper, we propose a minimal dialogue task which requires advanced skills of common grounding under continuous and partially-observable context. Based on this task formulation, we collected a largescale dataset of 6,760 dialogues which fulfills essential requirements of natural language corpora. Our analysis of the dataset revealed important phenomena related to common grounding that need to be considered. Finally, we evaluate and analyze baseline neural models on a simple subtask that requires recognition of the created common ground. We show that simple baseline models perform decently but leave room for further improvement. Overall, we show that our proposed task will be a fundamental testbed where we can train, evaluate, and analyze dialogue system’s ability for sophisticated common grounding.

Downloads

Published

2019-07-17

How to Cite

Udagawa, T., & Aizawa, A. (2019). A Natural Language Corpus of Common Grounding under Continuous and Partially-Observable Context. Proceedings of the AAAI Conference on Artificial Intelligence, 33(01), 7120-7127. https://doi.org/10.1609/aaai.v33i01.33017120

Issue

Section

AAAI Technical Track: Natural Language Processing