Manual data annotation is often slow, expensive, and difficult to scale-especially for tasks that are subjective and context-dependent, like detecting hate speech and stereotypes. With recent progress in Large Language Models (LLMs), there is growing potential to automate this process. In this study, first, we explore the use of a committee of LLMs (GPT-4omini, Gemini-1.5-flash, and DeepSeek-R1) to generate annotations for Italian social media content. Then, the quality of LLM-generated labels is evaluated by following a teacher-student approach, where we trained a smaller student model (phi3.5-miniinstruct) on them and testing its performance against a humanlabeled dataset. The results indicate that fine-tuning the student model over LLM-generated labels, with careful dataset balancing and hyperparameter tuning, can yield a significant improvement over the baseline (approximately 17% improvement), suggesting that committee-based LLM annotation can provide high-quality labels. Therefore, this work shows the proposed approach can be a reliable and scalable alternative to manual labeling by properly addressing challenges like class imbalance to be sure about the fairness and accuracy of the fine-tuned models.
From Annotation to Detection: Evaluating LLMs for Hate Speech and Stereotype Identification in Italian / F. Mohammadi, S. Maghool, P. Ceravolo - In: AICCSA[s.l] : Institute of Electrical and Electronics Engineers (IEEE), 2026 Jan 05. - ISBN 979-8-3315-5693-8. - pp. 1-6 (( 22. ACS International Conference on Computer Systems and Applications Doha (Qatar) 2025 [10.1109/aiccsa66935.2025.11315321].
From Annotation to Detection: Evaluating LLMs for Hate Speech and Stereotype Identification in Italian
F. Mohammadi
Primo
;S. MaghoolPenultimo
;P. CeravoloUltimo
2026
Abstract
Manual data annotation is often slow, expensive, and difficult to scale-especially for tasks that are subjective and context-dependent, like detecting hate speech and stereotypes. With recent progress in Large Language Models (LLMs), there is growing potential to automate this process. In this study, first, we explore the use of a committee of LLMs (GPT-4omini, Gemini-1.5-flash, and DeepSeek-R1) to generate annotations for Italian social media content. Then, the quality of LLM-generated labels is evaluated by following a teacher-student approach, where we trained a smaller student model (phi3.5-miniinstruct) on them and testing its performance against a humanlabeled dataset. The results indicate that fine-tuning the student model over LLM-generated labels, with careful dataset balancing and hyperparameter tuning, can yield a significant improvement over the baseline (approximately 17% improvement), suggesting that committee-based LLM annotation can provide high-quality labels. Therefore, this work shows the proposed approach can be a reliable and scalable alternative to manual labeling by properly addressing challenges like class imbalance to be sure about the fairness and accuracy of the fine-tuned models.| File | Dimensione | Formato | |
|---|---|---|---|
|
From_Annotation_to_Detection_Evaluating_LLMs_for_Hate_Speech_and_Stereotype_Identification_in_Italian.pdf
accesso riservato
Tipologia:
Publisher's version/PDF
Licenza:
Nessuna licenza
Dimensione
715.88 kB
Formato
Adobe PDF
|
715.88 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.




