COVYT: Introducing the Coronavirus YouTube and TikTok speech dataset
featuring the same speakers with and without infection
- URL: http://arxiv.org/abs/2206.11045v1
- Date: Mon, 20 Jun 2022 16:26:51 GMT
- Title: COVYT: Introducing the Coronavirus YouTube and TikTok speech dataset
featuring the same speakers with and without infection
- Authors: Andreas Triantafyllopoulos, Anastasia Semertzidou, Meishu Song,
Florian B. Pokorny, Bj\"orn W. Schuller
- Abstract summary: We introduce the COVYT dataset -- a novel COVID-19 dataset collected from public sources containing more than 8 hours of speech from 65 speakers.
As compared to other existing COVID-19 sound datasets, the unique feature of the COVYT dataset is that it comprises both COVID-19 positive and negative samples from all 65 speakers.
- Score: 4.894353840908006
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: More than two years after its outbreak, the COVID-19 pandemic continues to
plague medical systems around the world, putting a strain on scarce resources,
and claiming human lives. From the very beginning, various AI-based COVID-19
detection and monitoring tools have been pursued in an attempt to stem the tide
of infections through timely diagnosis. In particular, computer audition has
been suggested as a non-invasive, cost-efficient, and eco-friendly alternative
for detecting COVID-19 infections through vocal sounds. However, like all AI
methods, also computer audition is heavily dependent on the quantity and
quality of available data, and large-scale COVID-19 sound datasets are
difficult to acquire -- amongst other reasons -- due to the sensitive nature of
such data. To that end, we introduce the COVYT dataset -- a novel COVID-19
dataset collected from public sources containing more than 8 hours of speech
from 65 speakers. As compared to other existing COVID-19 sound datasets, the
unique feature of the COVYT dataset is that it comprises both COVID-19 positive
and negative samples from all 65 speakers. We analyse the acoustic
manifestation of COVID-19 on the basis of these perfectly speaker
characteristic balanced `in-the-wild' data using interpretable audio
descriptors, and investigate several classification scenarios that shed light
into proper partitioning strategies for a fair speech-based COVID-19 detection.
Related papers
- Statistical Design and Analysis for Robust Machine Learning: A Case
Study from COVID-19 [45.216628450147034]
This paper rigorously assesses state-of-the-art machine learning techniques used to predict COVID-19 infection status based on vocal audio signals.
We provide guidelines on testing the performance of methods to classify COVID-19 infection status based on acoustic features.
arXiv Detail & Related papers (2022-12-15T13:50:13Z) - Developing a multi-variate prediction model for the detection of
COVID-19 from Crowd-sourced Respiratory Voice Data [0.0]
The novelty of this work is in the development of a deep learning model for the identification of COVID-19 patients from voice recordings.
We used the Cambridge University dataset consisting of 893 audio samples, crowd-sourced from 4352 participants that used a COVID-19 Sounds app.
Based on the voice data, we developed deep learning classification models to detect positive COVID-19 cases.
arXiv Detail & Related papers (2022-09-08T11:46:37Z) - Evaluating the COVID-19 Identification ResNet (CIdeR) on the INTERSPEECH
COVID-19 from Audio Challenges [59.78485839636553]
CIdeR is an end-to-end deep learning neural network originally designed to classify whether an individual is COVID-positive or COVID-negative.
We demonstrate the potential of CIdeR at binary COVID-19 diagnosis from both the COVID-19 Cough and Speech Sub-Challenges of INTERSPEECH 2021, ComParE and DiCOVA.
arXiv Detail & Related papers (2021-07-30T10:59:08Z) - Project Achoo: A Practical Model and Application for COVID-19 Detection
from Recordings of Breath, Voice, and Cough [55.45063681652457]
We propose a machine learning method to quickly triage COVID-19 using recordings made on consumer devices.
The approach combines signal processing methods with fine-tuned deep learning networks and provides methods for signal denoising, cough detection and classification.
We have also developed and deployed a mobile application that uses symptoms checker together with voice, breath and cough signals to detect COVID-19 infection.
arXiv Detail & Related papers (2021-07-12T08:07:56Z) - Sounds of COVID-19: exploring realistic performance of audio-based
digital testing [17.59710651224251]
In this paper, we explore the realistic performance of audio-based digital testing of COVID-19.
We collected a large crowdsourced respiratory audio dataset through a mobile app, alongside recent COVID-19 test result and symptoms intended as a ground truth.
The unbiased model takes features extracted from breathing, coughs, and voice signals as predictors and yields an AUC-ROC of 0.71 (95% CI: 0.65$-$0.77)
arXiv Detail & Related papers (2021-06-29T15:50:36Z) - COVIDx-US -- An open-access benchmark dataset of ultrasound imaging data
for AI-driven COVID-19 analytics [116.6248556979572]
COVIDx-US is an open-access benchmark dataset of COVID-19 related ultrasound imaging data.
It consists of 93 lung ultrasound videos and 10,774 processed images of patients infected with SARS-CoV-2 pneumonia, non-SARS-CoV-2 pneumonia, as well as healthy control cases.
arXiv Detail & Related papers (2021-03-18T03:31:33Z) - Virufy: A Multi-Branch Deep Learning Network for Automated Detection of
COVID-19 [1.9899603776429056]
Researchers have successfully presented models for detecting COVID-19 infection status using audio samples recorded in clinical settings.
We propose a multi-branch deep learning network that is trained and tested on crowdsourced data where most of the data has not been manually processed and cleaned.
arXiv Detail & Related papers (2021-03-02T15:31:09Z) - Detecting COVID-19 from Breathing and Coughing Sounds using Deep Neural
Networks [68.8204255655161]
We adapt an ensemble of Convolutional Neural Networks to classify if a speaker is infected with COVID-19 or not.
Ultimately, it achieves an Unweighted Average Recall (UAR) of 74.9%, or an Area Under ROC Curve (AUC) of 80.7% by ensembling neural networks.
arXiv Detail & Related papers (2020-12-29T01:14:17Z) - Audio, Speech, Language, & Signal Processing for COVID-19: A
Comprehensive Overview [0.0]
The Coronavirus (COVID-19) pandemic has been the research focus world-wide in the year 2020.
A major portion of COVID-19 symptoms are related to the functioning of the respiratory system.
This drives the research focus towards identifying the markers of COVID-19 in speech and other human generated audio signals.
arXiv Detail & Related papers (2020-11-29T21:33:59Z) - CovidDeep: SARS-CoV-2/COVID-19 Test Based on Wearable Medical Sensors
and Efficient Neural Networks [51.589769497681175]
The novel coronavirus (SARS-CoV-2) has led to a pandemic.
The current testing regime based on Reverse Transcription-Polymerase Chain Reaction for SARS-CoV-2 has been unable to keep up with testing demands.
We propose a framework called CovidDeep that combines efficient DNNs with commercially available WMSs for pervasive testing of the virus.
arXiv Detail & Related papers (2020-07-20T21:47:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.