This paper is a collaborative effort between Linguistics, Law, and Computer Science to evaluate stereotypes and biases in automated translation systems. We advocate gender-neutral translation as a means to promote gender inclusion and improve the objectivity of machine translation. Our approach focuses on identifying gender bias in English-to-Italian translations. First, we define gender bias following human rights law and linguistics literature. Then we proceed by identifying gender-specific terms such as she/lei and he/lui as key elements. We then evaluate the cosine similarity between these target terms and others in the dataset to reveal the model's perception of semantic relations. Using numerical features, we effectively evaluate the intensity and direction of the bias. Our findings provide tangible insights for developing and training gender-neutral translation algorithms.
Identifying Gender Stereotypes and Biases in Automated Translation from English to Italian using Similarity Networks / F. Mohammadi, M.A. Tamborini, P. Ceravolo, C. Nardocci, S. Maghool (CEUR WORKSHOP PROCEEDINGS). - In: EWAF 2024 : European Workshop on Algorithmic Fairness 2024 / [a cura di] M. Cerrato, A. Vallenas Coronel, P. Ahrweiler, M. Loi, M. Pechenizkiy, A. Tamò-Larrieux. - [s.l] : CEUR-WS, 2025 Feb 17. - pp. 1-26 (( 3. European Workshop on Algorithmic Fairness Mainz 2024.
Identifying Gender Stereotypes and Biases in Automated Translation from English to Italian using Similarity Networks
F. Mohammadi
;M.A. Tamborini;P. Ceravolo;C. Nardocci;S. Maghool
2025
Abstract
This paper is a collaborative effort between Linguistics, Law, and Computer Science to evaluate stereotypes and biases in automated translation systems. We advocate gender-neutral translation as a means to promote gender inclusion and improve the objectivity of machine translation. Our approach focuses on identifying gender bias in English-to-Italian translations. First, we define gender bias following human rights law and linguistics literature. Then we proceed by identifying gender-specific terms such as she/lei and he/lui as key elements. We then evaluate the cosine similarity between these target terms and others in the dataset to reveal the model's perception of semantic relations. Using numerical features, we effectively evaluate the intensity and direction of the bias. Our findings provide tangible insights for developing and training gender-neutral translation algorithms.| File | Dimensione | Formato | |
|---|---|---|---|
|
paper_54.pdf
accesso aperto
Tipologia:
Publisher's version/PDF
Licenza:
Creative commons
Dimensione
1.28 MB
Formato
Adobe PDF
|
1.28 MB | Adobe PDF | Visualizza/Apri |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.




