Language Model Priming for Cross-Lingual Event Extraction

Authors

  • Steven Fincke Information Sciences Institute University of Southern California
  • Shantanu Agarwal Information Sciences Institute University of Southern California
  • Scott Miller Information Sciences Institute University of Southern California
  • Elizabeth Boschee Information Sciences Institute University of Southern California

DOI:

https://doi.org/10.1609/aaai.v36i10.21307

Keywords:

Speech & Natural Language Processing (SNLP)

Abstract

We present a novel, language-agnostic approach to "priming" language models for the task of event extraction, providing particularly effective performance in low-resource and zero-shot cross-lingual settings. With priming, we augment the input to the transformer stack's language model differently depending on the question(s) being asked of the model at runtime. For instance, if the model is being asked to identify arguments for the trigger "protested", we will provide that trigger as part of the input to the language model, allowing it to produce different representations for candidate arguments than when it is asked about arguments for the trigger "arrest" elsewhere in the same sentence. We show that by enabling the language model to better compensate for the deficits of sparse and noisy training data, our approach improves both trigger and argument detection and classification significantly over the state of the art in a zero-shot cross-lingual setting.

Downloads

Published

2022-06-28

How to Cite

Fincke, S., Agarwal, S., Miller, S., & Boschee, E. (2022). Language Model Priming for Cross-Lingual Event Extraction. Proceedings of the AAAI Conference on Artificial Intelligence, 36(10), 10627-10635. https://doi.org/10.1609/aaai.v36i10.21307

Issue

Section

AAAI Technical Track on Speech and Natural Language Processing