L3-Net Deep Audio Embeddings to Improve COVID-19 Detection from
Smartphone Data
- URL: http://arxiv.org/abs/2205.07682v1
- Date: Mon, 16 May 2022 13:50:22 GMT
- Title: L3-Net Deep Audio Embeddings to Improve COVID-19 Detection from
Smartphone Data
- Authors: Mattia Giovanni Campana, Andrea Rovati, Franca Delmastro, Elena Pagani
- Abstract summary: We investigate the capabilities of the proposed deep embedding model L3-Net to automatically extract meaningful features from raw respiratory audio recordings.
Results show that the combination of L3-Net with hand-crafted features overcomes the performance of the other works of 28.57% in terms of AUC.
- Score: 5.505634045241288
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Smartphones and wearable devices, along with Artificial Intelligence, can
represent a game-changer in the pandemic control, by implementing low-cost and
pervasive solutions to recognize the development of new diseases at their early
stages and by potentially avoiding the rise of new outbreaks. Some recent works
show promise in detecting diagnostic signals of COVID-19 from voice and coughs
by using machine learning and hand-crafted acoustic features. In this paper, we
decided to investigate the capabilities of the recently proposed deep embedding
model L3-Net to automatically extract meaningful features from raw respiratory
audio recordings in order to improve the performances of standard machine
learning classifiers in discriminating between COVID-19 positive and negative
subjects from smartphone data. We evaluated the proposed model on 3 datasets,
comparing the obtained results with those of two reference works. Results show
that the combination of L3-Net with hand-crafted features overcomes the
performance of the other works of 28.57% in terms of AUC in a set of
subject-independent experiments. This result paves the way to further
investigation on different deep audio embeddings, also for the automatic
detection of different diseases.
Related papers
- Utilizing Machine Learning and 3D Neuroimaging to Predict Hearing Loss: A Comparative Analysis of Dimensionality Reduction and Regression Techniques [0.0]
We have explored machine learning approaches for predicting hearing loss thresholds on the brain's gray matter 3D images.
In the first phase, we used a 3D CNN model to reduce high-dimensional input into latent space.
In the second phase, we utilized this model to reduce input into rich features.
arXiv Detail & Related papers (2024-04-30T18:39:41Z) - MLCA-AVSR: Multi-Layer Cross Attention Fusion based Audio-Visual Speech Recognition [62.89464258519723]
We propose a multi-layer cross-attention fusion based AVSR approach that promotes representation of each modality by fusing them at different levels of audio/visual encoders.
Our proposed approach surpasses the first-place system, establishing a new SOTA cpCER of 29.13% on this dataset.
arXiv Detail & Related papers (2024-01-07T08:59:32Z) - An Explainable Deep Learning-Based Method For Schizophrenia Diagnosis Using Generative Data-Augmentation [0.3222802562733786]
We leverage a deep learning-based method for the automatic diagnosis of schizophrenia using EEG brain recordings.
This approach utilizes generative data augmentation, a powerful technique that enhances the accuracy of the diagnosis.
arXiv Detail & Related papers (2023-10-25T12:55:16Z) - Transfer Learning for the Efficient Detection of COVID-19 from
Smartphone Audio Data [6.18778092044887]
Disease detection from smartphone data represents an open research challenge in mobile health (m-health) systems.
We present the experimental evaluation of 3 different deep learning models, compared also with hand-crafted features.
We evaluate the memory footprints of the different models for their possible applications on commercial mobile devices.
arXiv Detail & Related papers (2023-07-06T13:19:27Z) - Augment and Criticize: Exploring Informative Samples for Semi-Supervised
Monocular 3D Object Detection [64.65563422852568]
We improve the challenging monocular 3D object detection problem with a general semi-supervised framework.
We introduce a novel, simple, yet effective Augment and Criticize' framework that explores abundant informative samples from unlabeled data.
The two new detectors, dubbed 3DSeMo_DLE and 3DSeMo_FLEX, achieve state-of-the-art results with remarkable improvements for over 3.5% AP_3D/BEV (Easy) on KITTI.
arXiv Detail & Related papers (2023-03-20T16:28:15Z) - Project Achoo: A Practical Model and Application for COVID-19 Detection
from Recordings of Breath, Voice, and Cough [55.45063681652457]
We propose a machine learning method to quickly triage COVID-19 using recordings made on consumer devices.
The approach combines signal processing methods with fine-tuned deep learning networks and provides methods for signal denoising, cough detection and classification.
We have also developed and deployed a mobile application that uses symptoms checker together with voice, breath and cough signals to detect COVID-19 infection.
arXiv Detail & Related papers (2021-07-12T08:07:56Z) - Virufy: A Multi-Branch Deep Learning Network for Automated Detection of
COVID-19 [1.9899603776429056]
Researchers have successfully presented models for detecting COVID-19 infection status using audio samples recorded in clinical settings.
We propose a multi-branch deep learning network that is trained and tested on crowdsourced data where most of the data has not been manually processed and cleaned.
arXiv Detail & Related papers (2021-03-02T15:31:09Z) - Automated Model Design and Benchmarking of 3D Deep Learning Models for
COVID-19 Detection with Chest CT Scans [72.04652116817238]
We propose a differentiable neural architecture search (DNAS) framework to automatically search for the 3D DL models for 3D chest CT scans classification.
We also exploit the Class Activation Mapping (CAM) technique on our models to provide the interpretability of the results.
arXiv Detail & Related papers (2021-01-14T03:45:01Z) - End-2-End COVID-19 Detection from Breath & Cough Audio [68.41471917650571]
We demonstrate the first attempt to diagnose COVID-19 using end-to-end deep learning from a crowd-sourced dataset of audio samples.
We introduce a novel modelling strategy using a custom deep neural network to diagnose COVID-19 from a joint breath and cough representation.
arXiv Detail & Related papers (2021-01-07T01:13:00Z) - CovidDeep: SARS-CoV-2/COVID-19 Test Based on Wearable Medical Sensors
and Efficient Neural Networks [51.589769497681175]
The novel coronavirus (SARS-CoV-2) has led to a pandemic.
The current testing regime based on Reverse Transcription-Polymerase Chain Reaction for SARS-CoV-2 has been unable to keep up with testing demands.
We propose a framework called CovidDeep that combines efficient DNNs with commercially available WMSs for pervasive testing of the virus.
arXiv Detail & Related papers (2020-07-20T21:47:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.