Benchmarking of eight recurrent neural network variants for breath phase
and adventitious sound detection on a self-developed open-access lung sound
database-HF_Lung_V1
- URL: http://arxiv.org/abs/2102.03049v1
- Date: Fri, 5 Feb 2021 08:21:28 GMT
- Title: Benchmarking of eight recurrent neural network variants for breath phase
and adventitious sound detection on a self-developed open-access lung sound
database-HF_Lung_V1
- Authors: Fu-Shun Hsu, Shang-Ran Huang, Chien-Wen Huang, Chao-Jung Huang,
Yuan-Ren Cheng, Chun-Chieh Chen, Jack Hsiao, Chung-Wei Chen, Li-Chin Chen,
Yen-Chun Lai, Bi-Fang Hsu, Nian-Jhen Lin, Wan-Lin Tsai, Yi-Lin Wu, Tzu-Ling
Tseng, Ching-Ting Tseng, Yi-Tsun Chen, Feipei Lai
- Abstract summary: A robust computerized respiratory sound analysis algorithm has not yet been validated in practical applications.
We developed a lung sound database (HF_Lung_V1) comprising 9,765 audio files of lung sounds.
The results revealed that these models exhibited adequate performance in lung sound analysis.
- Score: 0.35775620736088914
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A reliable, remote, and continuous real-time respiratory sound monitor with
automated respiratory sound analysis ability is urgently required in many
clinical scenarios-such as in monitoring disease progression of coronavirus
disease 2019-to replace conventional auscultation with a handheld stethoscope.
However, a robust computerized respiratory sound analysis algorithm has not yet
been validated in practical applications. In this study, we developed a lung
sound database (HF_Lung_V1) comprising 9,765 audio files of lung sounds
(duration of 15 s each), 34,095 inhalation labels, 18,349 exhalation labels,
13,883 continuous adventitious sound (CAS) labels (comprising 8,457 wheeze
labels, 686 stridor labels, and 4,740 rhonchi labels), and 15,606 discontinuous
adventitious sound labels (all crackles). We conducted benchmark tests for long
short-term memory (LSTM), gated recurrent unit (GRU), bidirectional LSTM
(BiLSTM), bidirectional GRU (BiGRU), convolutional neural network (CNN)-LSTM,
CNN-GRU, CNN-BiLSTM, and CNN-BiGRU models for breath phase detection and
adventitious sound detection. We also conducted a performance comparison
between the LSTM-based and GRU-based models, between unidirectional and
bidirectional models, and between models with and without a CNN. The results
revealed that these models exhibited adequate performance in lung sound
analysis. The GRU-based models outperformed, in terms of F1 scores and areas
under the receiver operating characteristic curves, the LSTM-based models in
most of the defined tasks. Furthermore, all bidirectional models outperformed
their unidirectional counterparts. Finally, the addition of a CNN improved the
accuracy of lung sound analysis, especially in the CAS detection tasks.
Related papers
- Self-Supervised Pretraining Improves Performance and Inference
Efficiency in Multiple Lung Ultrasound Interpretation Tasks [65.23740556896654]
We investigated whether self-supervised pretraining could produce a neural network feature extractor applicable to multiple classification tasks in lung ultrasound analysis.
When fine-tuning on three lung ultrasound tasks, pretrained models resulted in an improvement of the average across-task area under the receiver operating curve (AUC) by 0.032 and 0.061 on local and external test sets respectively.
arXiv Detail & Related papers (2023-09-05T21:36:42Z) - Generalizing electrocardiogram delineation: training convolutional
neural networks with synthetic data augmentation [63.51064808536065]
Existing databases for ECG delineation are small, being insufficient in size and in the array of pathological conditions they represent.
This article delves has two main contributions. First, a pseudo-synthetic data generation algorithm was developed, based in probabilistically composing ECG traces given "pools" of fundamental segments, as cropped from the original databases, and a set of rules for their arrangement into coherent synthetic traces.
Second, two novel segmentation-based loss functions have been developed, which attempt at enforcing the prediction of an exact number of independent structures and at producing closer segmentation boundaries by focusing on a reduced number of samples.
arXiv Detail & Related papers (2021-11-25T10:11:41Z) - Cough Detection Using Selected Informative Features from Audio Signals [24.829135966052142]
The models are trained by the dataset combined ESC-50 dataset with self-recorded cough recordings.
The best cough detection model realizes the accuracy, recall, precision and F1-score with 94.9%, 97.1%, 93.1% and 0.95 respectively.
arXiv Detail & Related papers (2021-08-07T23:05:18Z) - Improved Breath Phase and Continuous Adventitious Sound Detection in
Lung and Tracheal Sound Using Mixed Set Training and Domain Adaptation [2.405718960148456]
We build a tracheal sound database, HF_Tracheal_V1, containing 11107 of 15-second tracheal sound recordings, 23087 inhalation labels, 16728 exhalation labels, and 6874 CAS labels.
The tracheal sound in HF_Tracheal_V1 and the lung sound in HF_Lung_V2 were either combined or used alone to train the CNN-BiGRU models for respective lung and tracheal sound analysis.
arXiv Detail & Related papers (2021-07-09T06:04:18Z) - Quantification of pulmonary involvement in COVID-19 pneumonia by means
of a cascade oftwo U-nets: training and assessment on multipledatasets using
different annotation criteria [83.83783947027392]
This study aims at exploiting Artificial intelligence (AI) for the identification, segmentation and quantification of COVID-19 pulmonary lesions.
We developed an automated analysis pipeline, the LungQuant system, based on a cascade of two U-nets.
The accuracy in predicting the CT-Severity Score (CT-SS) of the LungQuant system has been also evaluated.
arXiv Detail & Related papers (2021-05-06T10:21:28Z) - Crackle Detection In Lung Sounds Using Transfer Learning And Multi-Input
Convolitional Neural Networks [26.399917342840265]
We use transfer learning to tackle the mismatch of the recording setup for crackle detection in lung sounds.
A single input convolutional neural network (CNN) model is pre-trained on a source domain using ICBHI 2017, the largest publicly available database of lung sounds.
The multi-input model is then fine-tuned on the target domain of our self-collected lung sound database for classifying crackles and normal lung sounds.
arXiv Detail & Related papers (2021-04-30T11:32:42Z) - Detecting COVID-19 from Breathing and Coughing Sounds using Deep Neural
Networks [68.8204255655161]
We adapt an ensemble of Convolutional Neural Networks to classify if a speaker is infected with COVID-19 or not.
Ultimately, it achieves an Unweighted Average Recall (UAR) of 74.9%, or an Area Under ROC Curve (AUC) of 80.7% by ensembling neural networks.
arXiv Detail & Related papers (2020-12-29T01:14:17Z) - Multilabel 12-Lead Electrocardiogram Classification Using Gradient
Boosting Tree Ensemble [64.29529357862955]
We build an algorithm using gradient boosted tree ensembles fitted on morphology and signal processing features to classify ECG diagnosis.
For each lead, we derive features from heart rate variability, PQRST template shape, and the full signal waveform.
We join the features of all 12 leads to fit an ensemble of gradient boosting decision trees to predict probabilities of ECG instances belonging to each class.
arXiv Detail & Related papers (2020-10-21T18:11:36Z) - CovidDeep: SARS-CoV-2/COVID-19 Test Based on Wearable Medical Sensors
and Efficient Neural Networks [51.589769497681175]
The novel coronavirus (SARS-CoV-2) has led to a pandemic.
The current testing regime based on Reverse Transcription-Polymerase Chain Reaction for SARS-CoV-2 has been unable to keep up with testing demands.
We propose a framework called CovidDeep that combines efficient DNNs with commercially available WMSs for pervasive testing of the virus.
arXiv Detail & Related papers (2020-07-20T21:47:28Z) - A Systematic Search over Deep Convolutional Neural Network Architectures
for Screening Chest Radiographs [4.6411273009803065]
Chest radiographs are used for the screening of pulmonary and cardio-/thoracic conditions.
Recent efforts demonstrate a performance benchmark using an ensemble of deep convolutional neural networks (CNN)
Our systematic search over multiple standard CNN architectures identified single candidate models whose classification performances were found to be at par with ensembles.
arXiv Detail & Related papers (2020-04-24T12:30:40Z) - Deep Neural Network for Respiratory Sound Classification in Wearable
Devices Enabled by Patient Specific Model Tuning [2.8935588665357077]
We propose a deep CNN-RNN model that classifies respiratory sounds based on Mel-spectrograms.
We also implement a patient specific model tuning strategy that first screens respiratory patients and then builds patient specific classification models.
The proposed hybrid CNN-RNN model achieves a score of 66.31% on four-class classification of breathing cycles for ICBHI'17 scientific challenge respiratory sound database.
arXiv Detail & Related papers (2020-04-16T15:42:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.