Low Resource Quantitative Information Extraction via Structure Searching and Prefix-Based Text Generation

Authors

  • Tongliang Li Beihang University
  • Zixiang Wang Beihang University
  • Zhoujun Li Beihang University

DOI:

https://doi.org/10.1609/aaai.v37i11.26540

Keywords:

SNLP: Information Extraction, SNLP: Generation

Abstract

Quantitative information plays an important part in the financial and data analysis areas. Prior work relied on pattern-matching methods and complex hand-crafted rules to extract quantitative information due to the lack of labeled data. Such methods can be unstable and difficult to scale to the open domain. In this paper, we study quantitative information extraction in the low-resource setting. We propose a search-based approach by searching from the syntactic structures to acquire basic training data. The search process is simple yet effective. Then, a prefix-based text-to-text generation method is employed to extract the quantitative information. The prefix design can fully leverage pre-trained language models for text generation to serve the information extraction purpose. Experimental results show that our approaches achieves high performance with a limited amount of labeled data. The extraction result could further boost the performance of other tasks such as quantitative reasoning.

Downloads

Published

2023-06-26

How to Cite

Li, T., Wang, Z., & Li, Z. (2023). Low Resource Quantitative Information Extraction via Structure Searching and Prefix-Based Text Generation. Proceedings of the AAAI Conference on Artificial Intelligence, 37(11), 13112-13120. https://doi.org/10.1609/aaai.v37i11.26540

Issue

Section

AAAI Technical Track on Speech & Natural Language Processing