The ever-increasing interest in and widespread diffusion of Machine Learning (ML)-based applications has driven a substantial amount of research into offensive and defensive ML. ML models can be attacked from different angles: poisoning attacks, the focus of this paper, inject maliciously crafted data points in the training set to modify the model behavior; adversarial attacks maliciously manipulate inference-time data points to fool the ML model and drive the prediction of the ML model according to the attacker’s objective. Ensemble-based techniques are among the most relevant defenses against poisoning attacks and replace the monolithic ML model with an ensemble of ML models trained on different (disjoint) subsets of the training set. They assign data points to the training sets of the models in the ensemble (routing) randomly or using a hash function, assuming that evenly distributing poisoned data points positively influences ML robustness. Our paper departs from this assumption and implements a risk-based ensemble technique where a risk management process is used to perform a smart routing of data points to the training sets. An extensive experimental evaluation demonstrates the effectiveness of the proposed approach in terms of its soundness, robustness, and performance.

Protecting machine learning from poisoning attacks: A risk-based approach / N. Bena, M. Anisetti, E. Damiani, C.Y. Yeun, C.A. Ardagna. - In: COMPUTERS & SECURITY. - ISSN 0167-4048. - 155:(2025 Aug), pp. 104468.1-104468.13. [10.1016/j.cose.2025.104468]

Protecting machine learning from poisoning attacks: A risk-based approach

N. Bena
Primo
;
M. Anisetti
Secondo
;
E. Damiani;C.A. Ardagna
Ultimo
2025

Abstract

The ever-increasing interest in and widespread diffusion of Machine Learning (ML)-based applications has driven a substantial amount of research into offensive and defensive ML. ML models can be attacked from different angles: poisoning attacks, the focus of this paper, inject maliciously crafted data points in the training set to modify the model behavior; adversarial attacks maliciously manipulate inference-time data points to fool the ML model and drive the prediction of the ML model according to the attacker’s objective. Ensemble-based techniques are among the most relevant defenses against poisoning attacks and replace the monolithic ML model with an ensemble of ML models trained on different (disjoint) subsets of the training set. They assign data points to the training sets of the models in the ensemble (routing) randomly or using a hash function, assuming that evenly distributing poisoned data points positively influences ML robustness. Our paper departs from this assumption and implements a risk-based ensemble technique where a risk management process is used to perform a smart routing of data points to the training sets. An extensive experimental evaluation demonstrates the effectiveness of the proposed approach in terms of its soundness, robustness, and performance.
Ensemble; Machine learning; Poisoning; Risk; Robustness
Settore INFO-01/A - Informatica
   MUSA - Multilayered Urban Sustainability Actiona
   MUSA
   MINISTERO DELL'UNIVERSITA' E DELLA RICERCA

   BA-PHERD: Big Data Analytics Pipeline for the Identification of Heterogeneous Extracellular non-coding RNAs as Disease Biomarkers
   BA-PHERD
   MINISTERO DELL'UNIVERSITA' E DELLA RICERCA
   2022XABBMA_002

   SEcurity and RIghts in the CyberSpace (SERICS)
   SERICS
   MINISTERO DELL'UNIVERSITA' E DELLA RICERCA
   codice identificativo PE00000014

   One Health Action Hub: task force di Ateneo per la resilienza di ecosistemi territoriali (1H_Hub) - ONE HEALTH ACTION HUB
   (1H_Hub) - ONE HEALTH ACTION HUB
   UNIVERSITA' DEGLI STUDI DI MILANO

   Sovereign Edge-Hub: un’architettura cloud-edge per la sovranità digitale nelle scienze della vita - SOV-EDGE-HUB
   SOV-EDGE-HUB
   UNIVERSITA' DEGLI STUDI DI MILANO
ago-2025
Article (author)
File in questo prodotto:
File Dimensione Formato  
BADYA.COSE2025.pdf

accesso aperto

Tipologia: Publisher's version/PDF
Dimensione 2.21 MB
Formato Adobe PDF
2.21 MB Adobe PDF Visualizza/Apri
Pubblicazioni consigliate

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2434/1160875
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 1
  • ???jsp.display-item.citation.isi??? 0
  • OpenAlex ND
social impact