Quantum-Inspired Representation for Long-Tail Senses of Word Sense Disambiguation

Authors

  • Junwei Zhang Tianjin University
  • Ruifang He Tianjin University
  • Fengyu Guo Tianjin Normal University

DOI:

https://doi.org/10.1609/aaai.v37i11.26633

Keywords:

SNLP: Other Foundations of Speech & Natural Language Processing, ML: Quantum Machine Learning, ML: Representation Learning, SNLP: Language Grounding

Abstract

Data imbalance, also known as the long-tail distribution of data, is an important challenge for data-driven models. In the Word Sense Disambiguation (WSD) task, the long-tail phenomenon of word sense distribution is more common, making it difficult to effectively represent and identify Long-Tail Senses (LTSs). Therefore exploring representation methods that do not rely heavily on the training sample size is an important way to combat LTSs. Considering that many new states, namely superposition states, can be constructed from several known states in quantum mechanics, superposition states provide the possibility to obtain more accurate representations from inferior representations learned from a small sample size. Inspired by quantum superposition states, a representation method in Hilbert space is proposed to reduce the dependence on large sample sizes and thus combat LTSs. We theoretically prove the correctness of the method, and verify its effectiveness under the standard WSD evaluation framework and obtain state-of-the-art performance. Furthermore, we also test on the constructed LTS and the latest cross-lingual datasets, and achieve promising results.

Downloads

Published

2023-06-26

How to Cite

Zhang, J., He, R., & Guo, F. (2023). Quantum-Inspired Representation for Long-Tail Senses of Word Sense Disambiguation. Proceedings of the AAAI Conference on Artificial Intelligence, 37(11), 13949-13957. https://doi.org/10.1609/aaai.v37i11.26633

Issue

Section

AAAI Technical Track on Speech & Natural Language Processing