A large-scale and PCR-referenced vocal audio dataset for COVID-19
- URL: http://arxiv.org/abs/2212.07738v4
- Date: Fri, 3 Nov 2023 11:28:17 GMT
- Title: A large-scale and PCR-referenced vocal audio dataset for COVID-19
- Authors: Jobie Budd, Kieran Baker, Emma Karoune, Harry Coppock, Selina Patel,
Ana Tendero Ca\~nadas, Alexander Titcomb, Richard Payne, David Hurley,
Sabrina Egglestone, Lorraine Butler, Jonathon Mellor, George Nicholson, Ivan
Kiskin, Vasiliki Koutra, Radka Jersakova, Rachel A. McKendry, Peter Diggle,
Sylvia Richardson, Bj\"orn W. Schuller, Steven Gilmour, Davide Pigoli,
Stephen Roberts, Josef Packham, Tracey Thornley, Chris Holmes
- Abstract summary: The UK Health Security Agency recruited voluntary participants through the national Test and Trace programme and the REACT-1 survey in England from March 2021 to March 2022.
Audio recordings of influenzaal coughs, exhalations, and speech were collected in the 'Speak up to help beat coronavirus' digital survey.
This dataset has additional potential uses for bioacoustics research, with 11.30% participants reporting asthma.
- Score: 29.40538927182366
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The UK COVID-19 Vocal Audio Dataset is designed for the training and
evaluation of machine learning models that classify SARS-CoV-2 infection status
or associated respiratory symptoms using vocal audio. The UK Health Security
Agency recruited voluntary participants through the national Test and Trace
programme and the REACT-1 survey in England from March 2021 to March 2022,
during dominant transmission of the Alpha and Delta SARS-CoV-2 variants and
some Omicron variant sublineages. Audio recordings of volitional coughs,
exhalations, and speech were collected in the 'Speak up to help beat
coronavirus' digital survey alongside demographic, self-reported symptom and
respiratory condition data, and linked to SARS-CoV-2 test results. The UK
COVID-19 Vocal Audio Dataset represents the largest collection of SARS-CoV-2
PCR-referenced audio recordings to date. PCR results were linked to 70,794 of
72,999 participants and 24,155 of 25,776 positive cases. Respiratory symptoms
were reported by 45.62% of participants. This dataset has additional potential
uses for bioacoustics research, with 11.30% participants reporting asthma, and
27.20% with linked influenza PCR test results.
Related papers
- Coswara: A respiratory sounds and symptoms dataset for remote screening
of SARS-CoV-2 infection [23.789227109218118]
This paper presents the Coswara dataset, a dataset containing diverse set of respiratory sounds and rich meta-data.
The respiratory sounds contained nine sound categories associated with variants of breathing, cough and speech.
The paper summarizes the data collection procedure, demographic, symptoms and audio data information.
arXiv Detail & Related papers (2023-05-22T06:09:10Z) - Audio-based AI classifiers show no evidence of improved COVID-19
screening over simple symptoms checkers [37.085063562292845]
We collect and analyse a dataset of audio recordings from 67,842 individuals with linked metadata.
Subjects were recruited via the UK governments National Health Service Test-and-Trace programme and the REal-time Assessment of Community Transmission survey.
In an unadjusted analysis of our dataset AI classifiers predict SARS-CoV-2 infection status with high accuracy.
However, after matching on measured confounders, such as age, gender, and self reported symptoms, our classifiers performance is much weaker.
arXiv Detail & Related papers (2022-12-15T15:44:02Z) - Fruit-CoV: An Efficient Vision-based Framework for Speedy Detection and
Diagnosis of SARS-CoV-2 Infections Through Recorded Cough Sounds [0.38321248253111767]
SARS-CoV-2 has spread across the world, taking part in the global pandemic disease since March 2020.
It is vital to possess a self-testing service of SARS-CoV-2 at home.
In this study, we introduce Fruit-CoV, a two-stage vision framework, which is capable of detecting SARS-CoV-2 infections through recorded cough sounds.
arXiv Detail & Related papers (2021-09-06T07:56:02Z) - Evaluating the COVID-19 Identification ResNet (CIdeR) on the INTERSPEECH
COVID-19 from Audio Challenges [59.78485839636553]
CIdeR is an end-to-end deep learning neural network originally designed to classify whether an individual is COVID-positive or COVID-negative.
We demonstrate the potential of CIdeR at binary COVID-19 diagnosis from both the COVID-19 Cough and Speech Sub-Challenges of INTERSPEECH 2021, ComParE and DiCOVA.
arXiv Detail & Related papers (2021-07-30T10:59:08Z) - COVIDx-US -- An open-access benchmark dataset of ultrasound imaging data
for AI-driven COVID-19 analytics [116.6248556979572]
COVIDx-US is an open-access benchmark dataset of COVID-19 related ultrasound imaging data.
It consists of 93 lung ultrasound videos and 10,774 processed images of patients infected with SARS-CoV-2 pneumonia, non-SARS-CoV-2 pneumonia, as well as healthy control cases.
arXiv Detail & Related papers (2021-03-18T03:31:33Z) - Detecting COVID-19 from Breathing and Coughing Sounds using Deep Neural
Networks [68.8204255655161]
We adapt an ensemble of Convolutional Neural Networks to classify if a speaker is infected with COVID-19 or not.
Ultimately, it achieves an Unweighted Average Recall (UAR) of 74.9%, or an Area Under ROC Curve (AUC) of 80.7% by ensembling neural networks.
arXiv Detail & Related papers (2020-12-29T01:14:17Z) - Studying the Similarity of COVID-19 Sounds based on Correlation Analysis
of MFCC [1.9659095632676098]
We illustrate the importance of speech signal processing in the extraction of the Mel-Frequency Cepstral Coefficients (MFCCs) of the COVID-19 and non-COVID-19 samples.
Our results show high similarity in MFCCs between different COVID-19 cough and breathing sounds, while MFCC of voice is more robust between COVID-19 and non-COVID-19 samples.
arXiv Detail & Related papers (2020-10-17T11:38:05Z) - Pay Attention to the cough: Early Diagnosis of COVID-19 using
Interpretable Symptoms Embeddings with Cough Sound Signal Processing [0.0]
COVID-19 (coronavirus disease pandemic caused by SARS-CoV-2) has led to a treacherous and devastating catastrophe for humanity.
Current diagnosis of COVID-19 is done by Reverse-Transcription Polymer Chain Reaction (RT-PCR) testing.
An interpretable and COVID-19 diagnosis AI framework is devised and developed based on the cough sounds features and symptoms metadata.
arXiv Detail & Related papers (2020-10-06T01:22:50Z) - CovidDeep: SARS-CoV-2/COVID-19 Test Based on Wearable Medical Sensors
and Efficient Neural Networks [51.589769497681175]
The novel coronavirus (SARS-CoV-2) has led to a pandemic.
The current testing regime based on Reverse Transcription-Polymerase Chain Reaction for SARS-CoV-2 has been unable to keep up with testing demands.
We propose a framework called CovidDeep that combines efficient DNNs with commercially available WMSs for pervasive testing of the virus.
arXiv Detail & Related papers (2020-07-20T21:47:28Z) - COVID-Net S: Towards computer-aided severity assessment via training and
validation of deep neural networks for geographic extent and opacity extent
scoring of chest X-rays for SARS-CoV-2 lung disease severity [58.23203766439791]
Chest x-rays (CXRs) are often used to assess SARS-CoV-2 severity.
In this study, we assess the feasibility of computer-aided scoring of CXRs of SARS-CoV-2 lung disease severity using a deep learning system.
arXiv Detail & Related papers (2020-05-26T16:33:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.