Going Beyond the Cookie Theft Picture Test: Detecting Cognitive
Impairments using Acoustic Features
- URL: http://arxiv.org/abs/2206.05018v1
- Date: Fri, 10 Jun 2022 12:04:22 GMT
- Title: Going Beyond the Cookie Theft Picture Test: Detecting Cognitive
Impairments using Acoustic Features
- Authors: Franziska Braun, Andreas Erzigkeit, Hartmut Lehfeld, Thomas
Hillemacher, Korbinian Riedhammer, and Sebastian P. Bayerl
- Abstract summary: We show that acoustic features from standardized tests can be used to reliably discriminate cognitively impaired individuals from non-impaired ones.
We provide evidence that even features extracted from random speech samples of the interview can be a discriminator of cognitive impairment.
- Score: 0.18472148461613155
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Standardized tests play a crucial role in the detection of cognitive
impairment. Previous work demonstrated that automatic detection of cognitive
impairment is possible using audio data from a standardized picture description
task. The presented study goes beyond that, evaluating our methods on data
taken from two standardized neuropsychological tests, namely the German SKT and
a German version of the CERAD-NB, and a semi-structured clinical interview
between a patient and a psychologist. For the tests, we focus on speech
recordings of three sub-tests: reading numbers (SKT 3), interference (SKT 7),
and verbal fluency (CERAD-NB 1). We show that acoustic features from
standardized tests can be used to reliably discriminate cognitively impaired
individuals from non-impaired ones. Furthermore, we provide evidence that even
features extracted from random speech samples of the interview can be a
discriminator of cognitive impairment. In our baseline experiments, we use
OpenSMILE features and Support Vector Machine classifiers. In an improved
setup, we show that using wav2vec 2.0 features instead, we can achieve an
accuracy of up to 85%.
Related papers
- Infusing Acoustic Pause Context into Text-Based Dementia Assessment [7.8642589679025034]
This work investigates the use of pause-enriched transcripts in language models to differentiate the cognitive states of subjects with no cognitive impairment, mild cognitive impairment, and Alzheimer's dementia based on their speech from a clinical assessment.
The performance is evaluated through experiments on a German Verbal Fluency Test and a Picture Description Test, comparing the model's effectiveness across different speech production contexts.
arXiv Detail & Related papers (2024-08-27T16:44:41Z) - Identification of Cognitive Decline from Spoken Language through Feature
Selection and the Bag of Acoustic Words Model [0.0]
The early identification of symptoms of memory disorders plays a significant role in ensuring the well-being of populations.
The lack of standardized speech tests in clinical settings has led to a growing emphasis on developing automatic machine learning techniques for analyzing naturally spoken language.
The work presents an approach related to feature selection, allowing for the automatic selection of the essential features required for diagnosis from the Geneva minimalistic acoustic parameter set and relative speech pauses.
arXiv Detail & Related papers (2024-02-02T17:06:03Z) - PsyCoT: Psychological Questionnaire as Powerful Chain-of-Thought for
Personality Detection [50.66968526809069]
We propose a novel personality detection method, called PsyCoT, which mimics the way individuals complete psychological questionnaires in a multi-turn dialogue manner.
Our experiments demonstrate that PsyCoT significantly improves the performance and robustness of GPT-3.5 in personality detection.
arXiv Detail & Related papers (2023-10-31T08:23:33Z) - A New Benchmark of Aphasia Speech Recognition and Detection Based on
E-Branchformer and Multi-task Learning [29.916793641951507]
This paper presents a new benchmark for Aphasia speech recognition using state-of-the-art speech recognition techniques.
We introduce two multi-task learning methods based on the CTC/Attention architecture to perform both tasks simultaneously.
Our system achieves state-of-the-art speaker-level detection accuracy (97.3%), and a relative WER reduction of 11% for moderate Aphasia patients.
arXiv Detail & Related papers (2023-05-19T15:10:36Z) - Anomalous Sound Detection using Audio Representation with Machine ID
based Contrastive Learning Pretraining [52.191658157204856]
This paper uses contrastive learning to refine audio representations for each machine ID, rather than for each audio sample.
The proposed two-stage method uses contrastive learning to pretrain the audio representation model.
Experiments show that our method outperforms the state-of-the-art methods using contrastive learning or self-supervised classification.
arXiv Detail & Related papers (2023-04-07T11:08:31Z) - Automated Evaluation of Standardized Dementia Screening Tests [0.18472148461613155]
We report on a study that consists of a semi-standardized history taking followed by two standardized neuropsychological tests.
The tests include basic tasks such as naming objects, learning word lists, but also widely used tools such as the MMSE.
We show that using word alternatives helps to mitigate recognition errors and subsequently improves correlation with expert scores.
arXiv Detail & Related papers (2022-06-13T14:41:27Z) - Spotting adversarial samples for speaker verification by neural vocoders [102.1486475058963]
We adopt neural vocoders to spot adversarial samples for automatic speaker verification (ASV)
We find that the difference between the ASV scores for the original and re-synthesize audio is a good indicator for discrimination between genuine and adversarial samples.
Our codes will be made open-source for future works to do comparison.
arXiv Detail & Related papers (2021-07-01T08:58:16Z) - CogAlign: Learning to Align Textual Neural Representations to Cognitive
Language Processing Signals [60.921888445317705]
We propose a CogAlign approach to integrate cognitive language processing signals into natural language processing models.
We show that CogAlign achieves significant improvements with multiple cognitive features over state-of-the-art models on public datasets.
arXiv Detail & Related papers (2021-06-10T07:10:25Z) - NUVA: A Naming Utterance Verifier for Aphasia Treatment [49.114436579008476]
Assessment of speech performance using picture naming tasks is a key method for both diagnosis and monitoring of responses to treatment interventions by people with aphasia (PWA)
Here we present NUVA, an utterance verification system incorporating a deep learning element that classifies 'correct' versus'incorrect' naming attempts from aphasic stroke patients.
When tested on eight native British-English speaking PWA the system's performance accuracy ranged between 83.6% to 93.6%, with a 10-fold cross-validation mean of 89.5%.
arXiv Detail & Related papers (2021-02-10T13:00:29Z) - Speech Enhancement using Self-Adaptation and Multi-Head Self-Attention [70.82604384963679]
This paper investigates a self-adaptation method for speech enhancement using auxiliary speaker-aware features.
We extract a speaker representation used for adaptation directly from the test utterance.
arXiv Detail & Related papers (2020-02-14T05:05:36Z) - Accelerating Psychometric Screening Tests With Bayesian Active
Differential Selection [5.1779694507922835]
We propose a novel solution for rapid screening for a change in the psychometric function estimation of a given patient.
We validate our approach using audiometric data from the National Institute of Occupational Safety and Health NIOSH.
arXiv Detail & Related papers (2020-02-04T21:35:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.