This paper is a collaborative effort between Linguistics, Law, and Computer Science to evaluate stereotypes and biases in automated translation systems. We advocate gender-neutral translation as a means to promote gender inclusion and improve the objectivity of machine translation. Our approach focuses on identifying gender bias in English-to-Italian translations. First, we define gender bias following human rights law and linguistics literature. Then we proceed by identifying gender-specific terms such as she/lei and he/lui as key elements. We then evaluate the cosine similarity between these target terms and others in the dataset to reveal the model's perception of semantic relations. Using numerical features, we effectively evaluate the intensity and direction of the bias. Our findings provide tangible insights for developing and training gender-neutral translation algorithms.

Identifying Gender Stereotypes and Biases in Automated Translation from English to Italian using Similarity Networks / F. Mohammadi, M.A. Tamborini, P. Ceravolo, C. Nardocci, S. Maghool (CEUR WORKSHOP PROCEEDINGS). - In: EWAF 2024 : European Workshop on Algorithmic Fairness 2024 / [a cura di] M. Cerrato, A. Vallenas Coronel, P. Ahrweiler, M. Loi, M. Pechenizkiy, A. Tamò-Larrieux. - [s.l] : CEUR-WS, 2025 Feb 17. - pp. 1-26 (( 3. European Workshop on Algorithmic Fairness Mainz 2024.

Identifying Gender Stereotypes and Biases in Automated Translation from English to Italian using Similarity Networks

F. Mohammadi
;
M.A. Tamborini;P. Ceravolo;C. Nardocci;S. Maghool
2025

Abstract

This paper is a collaborative effort between Linguistics, Law, and Computer Science to evaluate stereotypes and biases in automated translation systems. We advocate gender-neutral translation as a means to promote gender inclusion and improve the objectivity of machine translation. Our approach focuses on identifying gender bias in English-to-Italian translations. First, we define gender bias following human rights law and linguistics literature. Then we proceed by identifying gender-specific terms such as she/lei and he/lui as key elements. We then evaluate the cosine similarity between these target terms and others in the dataset to reveal the model's perception of semantic relations. Using numerical features, we effectively evaluate the intensity and direction of the bias. Our findings provide tangible insights for developing and training gender-neutral translation algorithms.
Gender stereotypes; Gender bias; Large Language Models (LLMs); Automatic translation; Similarity networks; Inclusive language; Discrimination
Settore INFO-01/A - Informatica
17-feb-2025
https://ceur-ws.org/Vol-3908/paper_54.pdf
Book Part (author)
File in questo prodotto:
File Dimensione Formato  
paper_54.pdf

accesso aperto

Tipologia: Publisher's version/PDF
Licenza: Creative commons
Dimensione 1.28 MB
Formato Adobe PDF
1.28 MB Adobe PDF Visualizza/Apri
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2434/1157100
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus ND
  • ???jsp.display-item.citation.isi??? ND
  • OpenAlex ND
social impact