Detection of COVID-19 through the analysis of vocal fold oscillations
- URL: http://arxiv.org/abs/2010.10707v1
- Date: Wed, 21 Oct 2020 01:44:42 GMT
- Title: Detection of COVID-19 through the analysis of vocal fold oscillations
- Authors: Mahmoud Al Ismail, Soham Deshmukh, Rita Singh
- Abstract summary: Phonation, or the vibration of the vocal folds, is the primary source of vocalization in the production of voiced sounds by humans.
Since most symptomatic cases of COVID-19 present with moderate to severe impairment of respiratory functions, we hypothesize that signatures of COVID-19 may be observable by examining the vibrations of the vocal folds.
Our goal is to validate this hypothesis, and to quantitatively characterize the changes observed to enable the detection of COVID-19 from voice.
- Score: 18.387162887917164
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Phonation, or the vibration of the vocal folds, is the primary source of
vocalization in the production of voiced sounds by humans. It is a complex
bio-mechanical process that is highly sensitive to changes in the speaker's
respiratory parameters. Since most symptomatic cases of COVID-19 present with
moderate to severe impairment of respiratory functions, we hypothesize that
signatures of COVID-19 may be observable by examining the vibrations of the
vocal folds. Our goal is to validate this hypothesis, and to quantitatively
characterize the changes observed to enable the detection of COVID-19 from
voice. For this, we use a dynamical system model for the oscillation of the
vocal folds, and solve it using our recently developed ADLES algorithm to yield
vocal fold oscillation patterns directly from recorded speech. Experimental
results on a clinically curated dataset of COVID-19 positive and negative
subjects reveal characteristic patterns of vocal fold oscillations that are
correlated with COVID-19. We show that these are prominent and discriminative
enough that even simple classifiers such as logistic regression yields high
detection accuracies using just the recordings of isolated extended vowels.
Related papers
- Developing a multi-variate prediction model for the detection of
COVID-19 from Crowd-sourced Respiratory Voice Data [0.0]
The novelty of this work is in the development of a deep learning model for the identification of COVID-19 patients from voice recordings.
We used the Cambridge University dataset consisting of 893 audio samples, crowd-sourced from 4352 participants that used a COVID-19 Sounds app.
Based on the voice data, we developed deep learning classification models to detect positive COVID-19 cases.
arXiv Detail & Related papers (2022-09-08T11:46:37Z) - COVID-19 Detection from Respiratory Sounds with Hierarchical Spectrogram
Transformers [1.4091863292043447]
We introduce a novel deep learning approach to distinguish patients with COVID-19 from healthy controls given audio recordings of cough or breathing sounds.
The proposed approach leverages a novel hierarchical spectrogram transformer (HST) on spectrogram representations of respiratory sounds.
HST embodies self-attention mechanisms over local windows in spectrograms, and window size is progressively grown over model stages to capture local to global context.
arXiv Detail & Related papers (2022-07-19T19:55:16Z) - Evaluating the COVID-19 Identification ResNet (CIdeR) on the INTERSPEECH
COVID-19 from Audio Challenges [59.78485839636553]
CIdeR is an end-to-end deep learning neural network originally designed to classify whether an individual is COVID-positive or COVID-negative.
We demonstrate the potential of CIdeR at binary COVID-19 diagnosis from both the COVID-19 Cough and Speech Sub-Challenges of INTERSPEECH 2021, ComParE and DiCOVA.
arXiv Detail & Related papers (2021-07-30T10:59:08Z) - Project Achoo: A Practical Model and Application for COVID-19 Detection
from Recordings of Breath, Voice, and Cough [55.45063681652457]
We propose a machine learning method to quickly triage COVID-19 using recordings made on consumer devices.
The approach combines signal processing methods with fine-tuned deep learning networks and provides methods for signal denoising, cough detection and classification.
We have also developed and deployed a mobile application that uses symptoms checker together with voice, breath and cough signals to detect COVID-19 infection.
arXiv Detail & Related papers (2021-07-12T08:07:56Z) - DiffSVC: A Diffusion Probabilistic Model for Singing Voice Conversion [51.83469048737548]
We propose DiffSVC, an SVC system based on denoising diffusion probabilistic model.
A denoising module is trained in DiffSVC, which takes destroyed mel spectrogram and its corresponding step information as input to predict the added Gaussian noise.
Experiments show that DiffSVC can achieve superior conversion performance in terms of naturalness and voice similarity to current state-of-the-art SVC approaches.
arXiv Detail & Related papers (2021-05-28T14:26:40Z) - DiffSinger: Diffusion Acoustic Model for Singing Voice Synthesis [53.19363127760314]
DiffSinger is a parameterized Markov chain which iteratively converts the noise into mel-spectrogram conditioned on the music score.
The evaluations conducted on the Chinese singing dataset demonstrate that DiffSinger outperforms state-of-the-art SVS work with a notable margin.
arXiv Detail & Related papers (2021-05-06T05:21:42Z) - Detecting COVID-19 from Breathing and Coughing Sounds using Deep Neural
Networks [68.8204255655161]
We adapt an ensemble of Convolutional Neural Networks to classify if a speaker is infected with COVID-19 or not.
Ultimately, it achieves an Unweighted Average Recall (UAR) of 74.9%, or an Area Under ROC Curve (AUC) of 80.7% by ensembling neural networks.
arXiv Detail & Related papers (2020-12-29T01:14:17Z) - The voice of COVID-19: Acoustic correlates of infection [9.7390888107204]
COVID-19 is a global health crisis that has been affecting many aspects of our daily lives throughout the past year.
We compare acoustic features extracted from recordings of the vowels /i:/, /e:/, /o:/, /u:/, and /a:/ produced by 11 symptomatic COVID-19 positive and 11 COVID-19 negative German-speaking participants.
arXiv Detail & Related papers (2020-12-17T10:12:41Z) - Interpreting glottal flow dynamics for detecting COVID-19 from voice [18.387162887917164]
This paper proposes a method that analyzes the differential dynamics of the glottal flow waveform (GFW) during voice production.
We infer it from recorded speech signals and compare it to the GFW computed from physical model of phonation.
Our proposed method uses a CNN-based 2-step attention model that locates anomalies in time-feature space in the difference of the two GFWs.
arXiv Detail & Related papers (2020-10-29T13:16:57Z) - Studying the Similarity of COVID-19 Sounds based on Correlation Analysis
of MFCC [1.9659095632676098]
We illustrate the importance of speech signal processing in the extraction of the Mel-Frequency Cepstral Coefficients (MFCCs) of the COVID-19 and non-COVID-19 samples.
Our results show high similarity in MFCCs between different COVID-19 cough and breathing sounds, while MFCC of voice is more robust between COVID-19 and non-COVID-19 samples.
arXiv Detail & Related papers (2020-10-17T11:38:05Z) - Bulbar ALS Detection Based on Analysis of Voice Perturbation and Vibrato [68.97335984455059]
The purpose of this work was to verify the sutability of the sustain vowel phonation test for automatic detection of patients with ALS.
We proposed enhanced procedure for separation of voice signal into fundamental periods that requires for calculation of measurements.
arXiv Detail & Related papers (2020-03-24T12:49:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.