Identification of Cognitive Decline from Spoken Language through Feature
Selection and the Bag of Acoustic Words Model
- URL: http://arxiv.org/abs/2402.01824v1
- Date: Fri, 2 Feb 2024 17:06:03 GMT
- Title: Identification of Cognitive Decline from Spoken Language through Feature
Selection and the Bag of Acoustic Words Model
- Authors: Marko Niemel\"a and Mikaela von Bonsdorff and Sami \"Ayr\"am\"o and
Tommi K\"arkk\"ainen
- Abstract summary: The early identification of symptoms of memory disorders plays a significant role in ensuring the well-being of populations.
The lack of standardized speech tests in clinical settings has led to a growing emphasis on developing automatic machine learning techniques for analyzing naturally spoken language.
The work presents an approach related to feature selection, allowing for the automatic selection of the essential features required for diagnosis from the Geneva minimalistic acoustic parameter set and relative speech pauses.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Memory disorders are a central factor in the decline of functioning and daily
activities in elderly individuals. The confirmation of the illness, initiation
of medication to slow its progression, and the commencement of occupational
therapy aimed at maintaining and rehabilitating cognitive abilities require a
medical diagnosis. The early identification of symptoms of memory disorders,
especially the decline in cognitive abilities, plays a significant role in
ensuring the well-being of populations. Features related to speech production
are known to connect with the speaker's cognitive ability and changes. The lack
of standardized speech tests in clinical settings has led to a growing emphasis
on developing automatic machine learning techniques for analyzing naturally
spoken language. Non-lexical but acoustic properties of spoken language have
proven useful when fast, cost-effective, and scalable solutions are needed for
the rapid diagnosis of a disease. The work presents an approach related to
feature selection, allowing for the automatic selection of the essential
features required for diagnosis from the Geneva minimalistic acoustic parameter
set and relative speech pauses, intended for automatic paralinguistic and
clinical speech analysis. These features are refined into word histogram
features, in which machine learning classifiers are trained to classify control
subjects and dementia patients from the Dementia Bank's Pitt audio database.
The results show that achieving a 75% average classification accuracy with only
twenty-five features with the separate ADReSS 2020 competition test data and
the Leave-One-Subject-Out cross-validation of the entire competition data is
possible. The results rank at the top compared to international research, where
the same dataset and only acoustic features have been used to diagnose
patients.
Related papers
- Self-supervised Speech Models for Word-Level Stuttered Speech Detection [66.46810024006712]
We introduce a word-level stuttering speech detection model leveraging self-supervised speech models.
Our evaluation demonstrates that our model surpasses previous approaches in word-level stuttering speech detection.
arXiv Detail & Related papers (2024-09-16T20:18:20Z) - Selfsupervised learning for pathological speech detection [0.0]
Speech production is susceptible to influence and disruption by various neurodegenerative pathological speech disorders.
These disorders lead to pathological speech characterized by abnormal speech patterns and imprecise articulation.
Unlike neurotypical speakers, patients with speech pathologies or impairments are unable to access various virtual assistants such as Alexa, Siri, etc.
arXiv Detail & Related papers (2024-05-16T07:12:47Z) - Show from Tell: Audio-Visual Modelling in Clinical Settings [58.88175583465277]
We consider audio-visual modelling in a clinical setting, providing a solution to learn medical representations without human expert annotation.
A simple yet effective multi-modal self-supervised learning framework is proposed for this purpose.
The proposed approach is able to localise anatomical regions of interest during ultrasound imaging, with only speech audio as a reference.
arXiv Detail & Related papers (2023-10-25T08:55:48Z) - Leveraging Pretrained Representations with Task-related Keywords for
Alzheimer's Disease Detection [69.53626024091076]
Alzheimer's disease (AD) is particularly prominent in older adults.
Recent advances in pre-trained models motivate AD detection modeling to shift from low-level features to high-level representations.
This paper presents several efficient methods to extract better AD-related cues from high-level acoustic and linguistic features.
arXiv Detail & Related papers (2023-03-14T16:03:28Z) - Acoustic-Linguistic Features for Modeling Neurological Task Score in
Alzheimer's [1.290382979353427]
Natural language processing and machine learning provide promising techniques for reliably detecting Alzheimer's disease.
We compare and contrast the performance of ten linear regression models for predicting Mini-Mental Status exam scores.
We find that, for the given task, handcrafted linguistic features are more significant than acoustic and learned features.
arXiv Detail & Related papers (2022-09-13T15:35:31Z) - A Preliminary Study of a Two-Stage Paradigm for Preserving Speaker
Identity in Dysarthric Voice Conversion [50.040466658605524]
We propose a new paradigm for maintaining speaker identity in dysarthric voice conversion (DVC)
The poor quality of dysarthric speech can be greatly improved by statistical VC.
But as the normal speech utterances of a dysarthria patient are nearly impossible to collect, previous work failed to recover the individuality of the patient.
arXiv Detail & Related papers (2021-06-02T18:41:03Z) - NUVA: A Naming Utterance Verifier for Aphasia Treatment [49.114436579008476]
Assessment of speech performance using picture naming tasks is a key method for both diagnosis and monitoring of responses to treatment interventions by people with aphasia (PWA)
Here we present NUVA, an utterance verification system incorporating a deep learning element that classifies 'correct' versus'incorrect' naming attempts from aphasic stroke patients.
When tested on eight native British-English speaking PWA the system's performance accuracy ranged between 83.6% to 93.6%, with a 10-fold cross-validation mean of 89.5%.
arXiv Detail & Related papers (2021-02-10T13:00:29Z) - Predicting Early Indicators of Cognitive Decline from Verbal Utterances [2.387625146176821]
Dementia is a group of irreversible, chronic, and progressive neurodegenerative disorders resulting in impaired memory, communication, and thought processes.
We measure the feasibility of using the linguistic characteristics of verbal utterances elicited during neuropsychological exams to distinguish between elderly control groups, people with MCI, people diagnosed with possible Alzheimer's disease (AD), and probable AD.
Our experiments show that a combination of contextual and psycholinguistic features extracted by a Support Vector Machine improved distinguishing the verbal utterances of elderly controls, people with MCI, possible AD, and probable AD.
arXiv Detail & Related papers (2020-11-19T02:24:11Z) - Comparison of Speaker Role Recognition and Speaker Enrollment Protocol
for conversational Clinical Interviews [9.728371067160941]
We train end-to-end neural network architectures to adapt to each task and evaluate each approach under the same metric.
Results do not depend on the demographics of the Interviewee, highlighting the clinical relevance of our methods.
arXiv Detail & Related papers (2020-10-30T09:07:37Z) - Identification of Dementia Using Audio Biomarkers [15.740689461116762]
The objective of this work is to use speech processing and machine learning techniques to automatically identify the stage of dementia.
Non-linguistic acoustic parameters are used for this purpose, making this a language independent approach.
We analyze the contribution of various types of acoustic features such as spectral, temporal, cepstral their feature-level fusion and selection towards the identification of dementia stage.
arXiv Detail & Related papers (2020-02-27T13:54:00Z) - Speech Enhancement using Self-Adaptation and Multi-Head Self-Attention [70.82604384963679]
This paper investigates a self-adaptation method for speech enhancement using auxiliary speaker-aware features.
We extract a speaker representation used for adaptation directly from the test utterance.
arXiv Detail & Related papers (2020-02-14T05:05:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.