DA-Net: A Disentangled and Adaptive Network for Multi-Source Cross-Lingual Transfer Learning

Authors

  • Ling Ge School of Computer Science and Engineering, Beihang University, Beijing, China
  • Chunming Hu School of Computer Science and Engineering, Beihang University, Beijing, China College of Software, Beihang University, Beijing, China Zhongguancun Laboratory, Beijing, China
  • Guanghui Ma School of Computer Science and Engineering, Beihang University, Beijing, China
  • Jihong Liu School of Mechanical Engineering and Automation, Beihang University, Beijing, China
  • Hong Zhang National Computer Network Emergency Response Technical Team / Coordination Center of China, Beijing, China

DOI:

https://doi.org/10.1609/aaai.v38i16.29761

Keywords:

NLP: Machine Translation, Multilinguality, Cross-Lingual NLP, NLP: Applications

Abstract

Multi-Source cross-lingual transfer learning deals with the transfer of task knowledge from multiple labelled source languages to an unlabeled target language under the language shift. Existing methods typically focus on weighting the predictions produced by language-specific classifiers of different sources that follow a shared encoder. However, all source languages share the same encoder, which is updated by all these languages. The extracted representations inevitably contain different source languages' information, which may disturb the learning of the language-specific classifiers. Additionally, due to the language gap, language-specific classifiers trained with source labels are unable to make accurate predictions for the target language. Both facts impair the model's performance. To address these challenges, we propose a Disentangled and Adaptive Network ~(DA-Net). Firstly, we devise a feedback-guided collaborative disentanglement method that seeks to purify input representations of classifiers, thereby mitigating mutual interference from multiple sources. Secondly, we propose a class-aware parallel adaptation method that aligns class-level distributions for each source-target language pair, thereby alleviating the language pairs' language gap. Experimental results on three different tasks involving 38 languages validate the effectiveness of our approach.

Published

2024-03-24

How to Cite

Ge, L., Hu, C., Ma, G., Liu, J., & Zhang, H. (2024). DA-Net: A Disentangled and Adaptive Network for Multi-Source Cross-Lingual Transfer Learning. Proceedings of the AAAI Conference on Artificial Intelligence, 38(16), 18047-18055. https://doi.org/10.1609/aaai.v38i16.29761

Issue

Section

AAAI Technical Track on Natural Language Processing I