Linguistic-Based Mild Cognitive Impairment Detection Using Informative
Loss
- URL: http://arxiv.org/abs/2402.01690v1
- Date: Tue, 23 Jan 2024 16:30:22 GMT
- Title: Linguistic-Based Mild Cognitive Impairment Detection Using Informative
Loss
- Authors: Ali Pourramezan Fard, Mohammad H. Mahoor, Muath Alsuhaibani and Hiroko
H. Dodgec
- Abstract summary: We propose a framework that analyzes transcripts generated from video interviews collected within the I-CONECT study project.
Our framework can distinguish between MCI and NC with an average area under the curve of 84.75%.
- Score: 2.8893654860442872
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: This paper presents a deep learning method using Natural Language Processing
(NLP) techniques, to distinguish between Mild Cognitive Impairment (MCI) and
Normal Cognitive (NC) conditions in older adults. We propose a framework that
analyzes transcripts generated from video interviews collected within the
I-CONECT study project, a randomized controlled trial aimed at improving
cognitive functions through video chats. Our proposed NLP framework consists of
two Transformer-based modules, namely Sentence Embedding (SE) and Sentence
Cross Attention (SCA). First, the SE module captures contextual relationships
between words within each sentence. Subsequently, the SCA module extracts
temporal features from a sequence of sentences. This feature is then used by a
Multi-Layer Perceptron (MLP) for the classification of subjects into MCI or NC.
To build a robust model, we propose a novel loss function, called InfoLoss,
that considers the reduction in entropy by observing each sequence of sentences
to ultimately enhance the classification accuracy. The results of our
comprehensive model evaluation using the I-CONECT dataset show that our
framework can distinguish between MCI and NC with an average area under the
curve of 84.75%.
Related papers
- Continuous Sign Language Recognition Using Intra-inter Gloss Attention [0.0]
In this study, we introduce a novel module in sign language recognition studies, called intra-inter gloss attention module.
In the intra-gloss attention module, the video is divided into equally sized chunks and a self-attention mechanism is applied within each chunk.
Experimental results on the PHOENIX-2014 benchmark dataset demonstrate that our method can effectively extract sign language features in an end-to-end manner.
arXiv Detail & Related papers (2024-06-26T13:21:08Z) - CELA: Cost-Efficient Language Model Alignment for CTR Prediction [71.85120354973073]
Click-Through Rate (CTR) prediction holds a paramount position in recommender systems.
Recent efforts have sought to mitigate these challenges by integrating Pre-trained Language Models (PLMs)
We propose textbfCost-textbfEfficient textbfLanguage Model textbfAlignment (textbfCELA) for CTR prediction.
arXiv Detail & Related papers (2024-05-17T07:43:25Z) - FLIP: Fine-grained Alignment between ID-based Models and Pretrained Language Models for CTR Prediction [49.510163437116645]
Click-through rate (CTR) prediction plays as a core function module in personalized online services.
Traditional ID-based models for CTR prediction take as inputs the one-hot encoded ID features of tabular modality.
Pretrained Language Models(PLMs) has given rise to another paradigm, which takes as inputs the sentences of textual modality.
We propose to conduct Fine-grained feature-level ALignment between ID-based Models and Pretrained Language Models(FLIP) for CTR prediction.
arXiv Detail & Related papers (2023-10-30T11:25:03Z) - R-Cut: Enhancing Explainability in Vision Transformers with Relationship
Weighted Out and Cut [14.382326829600283]
We introduce two modules: the Relationship Weighted Out" and the Cut" modules.
The Cut" module performs fine-grained feature decomposition, taking into account factors such as position, texture, and color.
We validate our method with extensive qualitative and quantitative experiments on the ImageNet dataset.
arXiv Detail & Related papers (2023-07-18T08:03:51Z) - Scalable Learning of Latent Language Structure With Logical Offline
Cycle Consistency [71.42261918225773]
Conceptually, LOCCO can be viewed as a form of self-learning where the semantic being trained is used to generate annotations for unlabeled text.
As an added bonus, the annotations produced by LOCCO can be trivially repurposed to train a neural text generation model.
arXiv Detail & Related papers (2023-05-31T16:47:20Z) - Discover, Explanation, Improvement: An Automatic Slice Detection
Framework for Natural Language Processing [72.14557106085284]
slice detection models (SDM) automatically identify underperforming groups of datapoints.
This paper proposes a benchmark named "Discover, Explain, improve (DEIM)" for classification NLP tasks.
Our evaluation shows that Edisa can accurately select error-prone datapoints with informative semantic features.
arXiv Detail & Related papers (2022-11-08T19:00:00Z) - A Unified Understanding of Deep NLP Models for Text Classification [88.35418976241057]
We have developed a visual analysis tool, DeepNLPVis, to enable a unified understanding of NLP models for text classification.
The key idea is a mutual information-based measure, which provides quantitative explanations on how each layer of a model maintains the information of input words in a sample.
A multi-level visualization, which consists of a corpus-level, a sample-level, and a word-level visualization, supports the analysis from the overall training set to individual samples.
arXiv Detail & Related papers (2022-06-19T08:55:07Z) - Making a (Counterfactual) Difference One Rationale at a Time [5.97507595130844]
We investigate whether counterfactual data augmentation, without human assistance, can improve the performance of the selector.
Our results show that CDA produces rationales that better capture the signal of interest.
arXiv Detail & Related papers (2022-01-13T19:05:02Z) - SLM: Learning a Discourse Language Representation with Sentence
Unshuffling [53.42814722621715]
We introduce Sentence-level Language Modeling, a new pre-training objective for learning a discourse language representation.
We show that this feature of our model improves the performance of the original BERT by large margins.
arXiv Detail & Related papers (2020-10-30T13:33:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.