Interpreting Gender Bias in Neural Machine Translation: Multilingual Architecture Matters
DOI:
https://doi.org/10.1609/aaai.v36i11.21442Keywords:
AI For Social Impact (AISI Track Papers Only)Abstract
Multilingual neural machine translation architectures mainly differ in the number of sharing modules and parameters applied among languages. In this paper, and from an algorithmic perspective, we explore whether the chosen architecture, when trained with the same data, influences the level of gender bias. Experiments conducted in three language pairs show that language-specific encoder-decoders exhibit less bias than the shared architecture. We propose two methods for interpreting and studying gender bias in machine translation based on source embeddings and attention. Our analysis shows that, in the language-specific case, the embeddings encode more gender information, and their attention is more diverted. Both behaviors help in mitigating gender bias.Downloads
Published
2022-06-28
How to Cite
Costa-jussà, M. R., Escolano, C., Basta, C., Ferrando, J., Batlle, R., & Kharitonova, K. (2022). Interpreting Gender Bias in Neural Machine Translation: Multilingual Architecture Matters. Proceedings of the AAAI Conference on Artificial Intelligence, 36(11), 11855-11863. https://doi.org/10.1609/aaai.v36i11.21442
Issue
Section
AAAI Special Track on AI for Social Impact