Today’s society is witnessing not only an evergrowing depen- dency on data, but also an increasingly pervasiveness of related analytics and machine learning applications. From business to leisure, the avail- ability of services providing answers to questions brings great benefits in diverse domains. On the other side of the coin, the need to provide input data that the services need to compute a response. However, some data may be considered sensitive or confidential and users would legitimately be reluctant to release them to third parties. Considering classification tasks in machine learning applications, we in- troduce our PriSM (Privacy-friendly Support vector Machine) approach for computing a privacy-friendly model. PriSM anticipates the training phase of the classifier with a phase for discovering correlations among at- tributes that can indirectly expose sensitive information. It then trains the classifier excluding from consideration not only sensitive attributes but also other sets of attributes that have been learned as correlated to them. The result is a privacy-friendly classifier that does not require any of such information as input from the users. Our experimental evaluation on both synthetic and real-world datasets confirms the effectiveness of PriSM in protecting privacy while maintaining classification accuracy.
PriSM: A Privacy-Friendly Support Vector Machine / M. Barbato, A. Ceselli, S. De Capitani Di Vimercati, S. Foresti, P. Samarati (LECTURE NOTES IN COMPUTER SCIENCE). - In: Computer Security - ESORICS 2025 / [a cura di] V. Nicomette, A. Benzekri, N. Boulahia-Cuppens, J. Vaidya. - [s.l] : Springer, 2025. - ISBN 9783032078834. - pp. 62-82 (( Intervento presentato al 30. convegno European Symposium on Research in Computer Security ( Part 1) : September 22–24 tenutosi a Toulouse nel 2025 [10.1007/978-3-032-07884-1_4].
PriSM: A Privacy-Friendly Support Vector Machine
M. Barbato;A. Ceselli;S. De Capitani Di Vimercati;S. Foresti;P. Samarati
2025
Abstract
Today’s society is witnessing not only an evergrowing depen- dency on data, but also an increasingly pervasiveness of related analytics and machine learning applications. From business to leisure, the avail- ability of services providing answers to questions brings great benefits in diverse domains. On the other side of the coin, the need to provide input data that the services need to compute a response. However, some data may be considered sensitive or confidential and users would legitimately be reluctant to release them to third parties. Considering classification tasks in machine learning applications, we in- troduce our PriSM (Privacy-friendly Support vector Machine) approach for computing a privacy-friendly model. PriSM anticipates the training phase of the classifier with a phase for discovering correlations among at- tributes that can indirectly expose sensitive information. It then trains the classifier excluding from consideration not only sensitive attributes but also other sets of attributes that have been learned as correlated to them. The result is a privacy-friendly classifier that does not require any of such information as input from the users. Our experimental evaluation on both synthetic and real-world datasets confirms the effectiveness of PriSM in protecting privacy while maintaining classification accuracy.| File | Dimensione | Formato | |
|---|---|---|---|
|
esorics2025.pdf
accesso riservato
Tipologia:
Publisher's version/PDF
Licenza:
Nessuna licenza
Dimensione
1.07 MB
Formato
Adobe PDF
|
1.07 MB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.




