Smartphones and wearable devices, along with Artificial Intelligence, can represent a game-changer in the pandemic control, by implementing low-cost and pervasive solutions to recognize the development of new diseases at their early stages and by potentially avoiding the rise of new outbreaks. Some recent works show promise in detecting diagnostic signals of COVID-19 from voice and coughs by using machine learning and hand-crafted acoustic features. In this paper, we decided to investigate the capabilities of the recently proposed deep embedding model L-3-Net to automatically extract meaningful features from raw respiratory audio recordings in order to improve the performances of standard machine learning classifiers in discriminating between COVID-19 positive and negative subjects from smartphone data. We evaluated the proposed model on 3 datasets, comparing the obtained results with those of two reference works. Results show that the combination of L-3-Net with hand-crafted features overcomes the performance of the other works of 28.57% in terms of AUC in a set of subject-independent experiments. This result paves the way to further investigation on different deep audio embeddings, also for the automatic detection of different diseases.
L3-Net Deep Audio Embeddings to Improve COVID-19 Detection from Smartphone Data / M. Campana, A. Rovati, F. Delmastro, E. Pagani - In: 2022 IEEE International Conference on Smart Computing (SMARTCOMP)[s.l] : IEEE, 2022. - ISBN 978-1-6654-8152-6. - pp. 100-107 (( Intervento presentato al 8. convegno International Conference on Smart Computing (SMARTCOMP) tenutosi a Espoo nel 2022 [10.1109/SMARTCOMP55677.2022.00029].
L3-Net Deep Audio Embeddings to Improve COVID-19 Detection from Smartphone Data
E. PaganiUltimo
2022
Abstract
Smartphones and wearable devices, along with Artificial Intelligence, can represent a game-changer in the pandemic control, by implementing low-cost and pervasive solutions to recognize the development of new diseases at their early stages and by potentially avoiding the rise of new outbreaks. Some recent works show promise in detecting diagnostic signals of COVID-19 from voice and coughs by using machine learning and hand-crafted acoustic features. In this paper, we decided to investigate the capabilities of the recently proposed deep embedding model L-3-Net to automatically extract meaningful features from raw respiratory audio recordings in order to improve the performances of standard machine learning classifiers in discriminating between COVID-19 positive and negative subjects from smartphone data. We evaluated the proposed model on 3 datasets, comparing the obtained results with those of two reference works. Results show that the combination of L-3-Net with hand-crafted features overcomes the performance of the other works of 28.57% in terms of AUC in a set of subject-independent experiments. This result paves the way to further investigation on different deep audio embeddings, also for the automatic detection of different diseases.File | Dimensione | Formato | |
---|---|---|---|
2205.07682.pdf
accesso aperto
Tipologia:
Pre-print (manoscritto inviato all'editore)
Dimensione
834.93 kB
Formato
Adobe PDF
|
834.93 kB | Adobe PDF | Visualizza/Apri |
L3-Net_Deep_Audio_Embeddings_to_Improve_COVID-19_Detection_from_Smartphone_Data.pdf
accesso riservato
Tipologia:
Publisher's version/PDF
Dimensione
490.88 kB
Formato
Adobe PDF
|
490.88 kB | Adobe PDF | Visualizza/Apri Richiedi una copia |
Pubblicazioni consigliate
I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.