Classification of ALS patients based on acoustic analysis of sustained
vowel phonations
- URL: http://arxiv.org/abs/2012.07347v2
- Date: Mon, 11 Jan 2021 08:41:07 GMT
- Title: Classification of ALS patients based on acoustic analysis of sustained
vowel phonations
- Authors: Maxim Vashkevich and Yulia Rushkevich
- Abstract summary: Amyotrophic lateral sclerosis (ALS) is incurable neurological disorder with rapidly progressive course.
Common early symptoms of ALS are difficulty in swallowing and speech.
This study presents an approach to voice assessment for automatic system that separates healthy people from patients with ALS.
- Score: 0.7106986689736825
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Amyotrophic lateral sclerosis (ALS) is incurable neurological disorder with
rapidly progressive course. Common early symptoms of ALS are difficulty in
swallowing and speech. However, early acoustic manifestation of speech and
voice symptoms is very variable, that making their detection very challenging,
both by human specialists and automatic systems. This study presents an
approach to voice assessment for automatic system that separates healthy people
from patients with ALS. In particular, this work focus on analysing of sustain
phonation of vowels /a/ and /i/ to perform automatic classification of ALS
patients. A wide range of acoustic features such as MFCC, formants, jitter,
shimmer, vibrato, PPE, GNE, HNR, etc. were analysed. We also proposed a new set
of acoustic features for characterizing harmonic structure of the vowels.
Calculation of these features is based on pitch synchronized voice analysis. A
linear discriminant analysis (LDA) was used to classify the phonation produced
by patients with ALS and those by healthy individuals. Several algorithms of
feature selection were tested to find optimal feature subset for LDA model. The
study's experiments show that the most successful LDA model based on 32
features picked out by LASSO feature selection algorithm attains 99.7% accuracy
with 99.3% sensitivity and 99.9% specificity. Among the classifiers with a
small number of features, we can highlight LDA model with 5 features, which has
89.0% accuracy (87.5% sensitivity and 90.4% specificity).
Related papers
- Self-supervised ASR Models and Features For Dysarthric and Elderly Speech Recognition [71.87998918300806]
This paper explores approaches to integrate domain fine-tuned SSL pre-trained models and their features into TDNN and Conformer ASR systems.
TDNN systems constructed by integrating domain-adapted HuBERT, wav2vec2-conformer or multi-lingual XLSR models consistently outperform standalone fine-tuned SSL pre-trained models.
Consistent improvements in Alzheimer's Disease detection accuracy are also obtained using the DementiaBank Pitt elderly speech recognition outputs.
arXiv Detail & Related papers (2024-07-03T08:33:39Z) - Automatic Prediction of Amyotrophic Lateral Sclerosis Progression using Longitudinal Speech Transformer [56.17737749551133]
We propose ALS longitudinal speech transformer (ALST), a neural network-based automatic predictor of ALS disease progression.
By taking advantage of high-quality pretrained speech features and longitudinal information in the recordings, our best model achieves 91.0% AUC.
ALST is capable of fine-grained and interpretable predictions of ALS progression, especially for distinguishing between rarer and more severe cases.
arXiv Detail & Related papers (2024-06-26T13:28:24Z) - Remote Inference of Cognitive Scores in ALS Patients Using a Picture
Description [3.441452604187627]
We implement the digital version of the Edinburgh Cognitive and Behavioral ALS Screen test for the first time.
This test which is designed to measure cognitive impairment was remotely performed by 56 participants from the EverythingALS Speech Study.
We analyze the descriptions performed within +/- 60 days from the day the ECAS test was administered and extract different types of linguistic and acoustic features.
arXiv Detail & Related papers (2023-09-13T14:30:30Z) - Acoustic-to-articulatory inversion for dysarthric speech: Are
pre-trained self-supervised representations favorable? [3.43759997215733]
Acoustic-to-articulatory inversion (AAI) involves mapping from the acoustic to the articulatory space.
In this work, we perform AAI for dysarthric speech using representations from pre-trained self-supervised learning (SSL) models.
arXiv Detail & Related papers (2023-09-03T07:44:38Z) - Exploiting prompt learning with pre-trained language models for
Alzheimer's Disease detection [70.86672569101536]
Early diagnosis of Alzheimer's disease (AD) is crucial in facilitating preventive care and to delay further progression.
This paper investigates the use of prompt-based fine-tuning of PLMs that consistently uses AD classification errors as the training objective function.
arXiv Detail & Related papers (2022-10-29T09:18:41Z) - Exploring linguistic feature and model combination for speech
recognition based automatic AD detection [61.91708957996086]
Speech based automatic AD screening systems provide a non-intrusive and more scalable alternative to other clinical screening techniques.
Scarcity of specialist data leads to uncertainty in both model selection and feature learning when developing such systems.
This paper investigates the use of feature and model combination approaches to improve the robustness of domain fine-tuning of BERT and Roberta pre-trained text encoders.
arXiv Detail & Related papers (2022-06-28T05:09:01Z) - Continuous Speech for Improved Learning Pathological Voice Disorders [12.867900671251395]
This study proposes a novel approach, using continuous Mandarin speech instead of a single vowel, to classify four common voice disorders.
In the proposed framework, acoustic signals are transformed into mel-frequency cepstral coefficients, and a bi-directional long-short term memory network (BiLSTM) is adopted to model the sequential features.
arXiv Detail & Related papers (2022-02-22T09:58:31Z) - NUVA: A Naming Utterance Verifier for Aphasia Treatment [49.114436579008476]
Assessment of speech performance using picture naming tasks is a key method for both diagnosis and monitoring of responses to treatment interventions by people with aphasia (PWA)
Here we present NUVA, an utterance verification system incorporating a deep learning element that classifies 'correct' versus'incorrect' naming attempts from aphasic stroke patients.
When tested on eight native British-English speaking PWA the system's performance accuracy ranged between 83.6% to 93.6%, with a 10-fold cross-validation mean of 89.5%.
arXiv Detail & Related papers (2021-02-10T13:00:29Z) - Bulbar ALS Detection Based on Analysis of Voice Perturbation and Vibrato [68.97335984455059]
The purpose of this work was to verify the sutability of the sustain vowel phonation test for automatic detection of patients with ALS.
We proposed enhanced procedure for separation of voice signal into fundamental periods that requires for calculation of measurements.
arXiv Detail & Related papers (2020-03-24T12:49:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.