Towards Interpretability of Speech Pause in Dementia Detection using
Adversarial Learning
- URL: http://arxiv.org/abs/2111.07454v1
- Date: Sun, 14 Nov 2021 21:26:18 GMT
- Title: Towards Interpretability of Speech Pause in Dementia Detection using
Adversarial Learning
- Authors: Youxiang Zhu, Bang Tran, Xiaohui Liang, John A. Batsis, Robert M. Roth
- Abstract summary: Speech pause is an effective biomarker in dementia detection.
Recent deep learning models have exploited speech pauses to achieve highly accurate dementia detection.
We will study the positions and lengths of dementia-sensitive pauses using adversarial learning approaches.
- Score: 4.19159477763309
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Speech pause is an effective biomarker in dementia detection. Recent deep
learning models have exploited speech pauses to achieve highly accurate
dementia detection, but have not exploited the interpretability of speech
pauses, i.e., what and how positions and lengths of speech pauses affect the
result of dementia detection. In this paper, we will study the positions and
lengths of dementia-sensitive pauses using adversarial learning approaches.
Specifically, we first utilize an adversarial attack approach by adding the
perturbation to the speech pauses of the testing samples, aiming to reduce the
confidence levels of the detection model. Then, we apply an adversarial
training approach to evaluate the impact of the perturbation in training
samples on the detection model. We examine the interpretability from the
perspectives of model accuracy, pause context, and pause length. We found that
some pauses are more sensitive to dementia than other pauses from the model's
perspective, e.g., speech pauses near to the verb "is". Increasing lengths of
sensitive pauses or adding sensitive pauses leads the model inference to
Alzheimer's Disease, while decreasing the lengths of sensitive pauses or
deleting sensitive pauses leads to non-AD.
Related papers
- Self-supervised Speech Models for Word-Level Stuttered Speech Detection [66.46810024006712]
We introduce a word-level stuttering speech detection model leveraging self-supervised speech models.
Our evaluation demonstrates that our model surpasses previous approaches in word-level stuttering speech detection.
arXiv Detail & Related papers (2024-09-16T20:18:20Z) - Infusing Acoustic Pause Context into Text-Based Dementia Assessment [7.8642589679025034]
This work investigates the use of pause-enriched transcripts in language models to differentiate the cognitive states of subjects with no cognitive impairment, mild cognitive impairment, and Alzheimer's dementia based on their speech from a clinical assessment.
The performance is evaluated through experiments on a German Verbal Fluency Test and a Picture Description Test, comparing the model's effectiveness across different speech production contexts.
arXiv Detail & Related papers (2024-08-27T16:44:41Z) - Inappropriate Pause Detection In Dysarthric Speech Using Large-Scale
Speech Recognition [2.7309692684728617]
Inappropriate pauses are crucial indicators in severity assessment and speech-language therapy.
We propose a large-scale speech recognition model for inappropriate pause detection in dysarthric speech.
Our experiments show that the proposed method better detects inappropriate pauses in dysarthric speech than baselines.
arXiv Detail & Related papers (2024-02-29T07:29:42Z) - Regressor-Segmenter Mutual Prompt Learning for Crowd Counting [70.49246560246736]
We propose mutual prompt learning (mPrompt) to solve bias and inaccuracy caused by annotation variance.
Experiments show that mPrompt significantly reduces the Mean Average Error (MAE)
arXiv Detail & Related papers (2023-12-04T07:53:59Z) - Automatically measuring speech fluency in people with aphasia: first
achievements using read-speech data [55.84746218227712]
This study aims at assessing the relevance of a signalprocessingalgorithm, initially developed in the field of language acquisition, for the automatic measurement of speech fluency.
arXiv Detail & Related papers (2023-08-09T07:51:40Z) - Leveraging Pretrained Representations with Task-related Keywords for
Alzheimer's Disease Detection [69.53626024091076]
Alzheimer's disease (AD) is particularly prominent in older adults.
Recent advances in pre-trained models motivate AD detection modeling to shift from low-level features to high-level representations.
This paper presents several efficient methods to extract better AD-related cues from high-level acoustic and linguistic features.
arXiv Detail & Related papers (2023-03-14T16:03:28Z) - Deep Learning for Hate Speech Detection: A Comparative Study [54.42226495344908]
We present here a large-scale empirical comparison of deep and shallow hate-speech detection methods.
Our goal is to illuminate progress in the area, and identify strengths and weaknesses in the current state-of-the-art.
In doing so we aim to provide guidance as to the use of hate-speech detection in practice, quantify the state-of-the-art, and identify future research directions.
arXiv Detail & Related papers (2022-02-19T03:48:20Z) - Incremental False Negative Detection for Contrastive Learning [95.68120675114878]
We introduce a novel incremental false negative detection for self-supervised contrastive learning.
During contrastive learning, we discuss two strategies to explicitly remove the detected false negatives.
Our proposed method outperforms other self-supervised contrastive learning frameworks on multiple benchmarks within a limited compute.
arXiv Detail & Related papers (2021-06-07T15:29:14Z) - Stuttering Speech Disfluency Prediction using Explainable Attribution
Vectors of Facial Muscle Movements [2.6540572249827514]
Speech disorders such as stuttering disrupt the normal fluency of speech by involuntary repetitions, prolongations and blocking of sounds and syllables.
Recent studies have explored automatic detection of stuttering using Artificial Intelligence (AI) based algorithm from respiratory rate, audio, etc. during speech utterance.
We hypothesize that pre-speech facial activity in AWS, which can be captured non-invasively, contains enough information to accurately classify the upcoming utterance as either fluent or stuttered.
arXiv Detail & Related papers (2020-10-02T23:45:41Z) - Identification of Dementia Using Audio Biomarkers [15.740689461116762]
The objective of this work is to use speech processing and machine learning techniques to automatically identify the stage of dementia.
Non-linguistic acoustic parameters are used for this purpose, making this a language independent approach.
We analyze the contribution of various types of acoustic features such as spectral, temporal, cepstral their feature-level fusion and selection towards the identification of dementia stage.
arXiv Detail & Related papers (2020-02-27T13:54:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.