Interpreting Gender Bias in Neural Machine Translation: Multilingual Architecture Matters
Marta R. Costa-jussà, Carlos C Escolano, Christine Raouf Saad Basta, Javier Ferrando Monsonís, Roser Batlle Roca, Ksenia Kharitonova
[AAAI-22] AI for Social Impact Track
Abstract:
Multilingual neural machine translation architectures mainly differ in the number of sharing modules and parameters applied among languages. In this paper, and from an algorithmic perspective, we explore whether the chosen architecture, when trained with the same data, influences the level of gender bias. Experiments conducted in three language pairs show that language-specific encoder-decoders exhibit less bias than the shared architecture. We propose two methods for interpreting and studying gender bias in machine translation based on source embeddings and attention. Our analysis shows that, in the language-specific case, the embeddings encode more gender information, and their attention is more diverted. Both behaviors help in mitigating gender bias.
Introduction Video
Sessions where this paper appears
-
Poster Session 3
Fri, February 25 8:45 AM - 10:30 AM (+00:00)
Red 6
-
Poster Session 12
Mon, February 28 8:45 AM - 10:30 AM (+00:00)
Red 6