A Depression Detection Method Based on Multi-Modal Feature Fusion Using Cross-Attention
- URL: http://arxiv.org/abs/2407.12825v1
- Date: Tue, 2 Jul 2024 13:13:35 GMT
- Title: A Depression Detection Method Based on Multi-Modal Feature Fusion Using Cross-Attention
- Authors: Shengjie Li, Yinhao Xiao,
- Abstract summary: Depression affects approximately 3.8% of the global population.
Over 75% of individuals in low- and middle-income countries remain untreated.
This paper introduces a novel method for detecting depression based on multi-modal feature fusion utilizing cross-attention.
- Score: 3.4872769952628926
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Depression, a prevalent and serious mental health issue, affects approximately 3.8\% of the global population. Despite the existence of effective treatments, over 75\% of individuals in low- and middle-income countries remain untreated, partly due to the challenge in accurately diagnosing depression in its early stages. This paper introduces a novel method for detecting depression based on multi-modal feature fusion utilizing cross-attention. By employing MacBERT as a pre-training model to extract lexical features from text and incorporating an additional Transformer module to refine task-specific contextual understanding, the model's adaptability to the targeted task is enhanced. Diverging from previous practices of simply concatenating multimodal features, this approach leverages cross-attention for feature integration, significantly improving the accuracy in depression detection and enabling a more comprehensive and precise analysis of user emotions and behaviors. Furthermore, a Multi-Modal Feature Fusion Network based on Cross-Attention (MFFNC) is constructed, demonstrating exceptional performance in the task of depression identification. The experimental results indicate that our method achieves an accuracy of 0.9495 on the test dataset, marking a substantial improvement over existing approaches. Moreover, it outlines a promising methodology for other social media platforms and tasks involving multi-modal processing. Timely identification and intervention for individuals with depression are crucial for saving lives, highlighting the immense potential of technology in facilitating early intervention for mental health issues.
Related papers
- Confidence-aware multi-modality learning for eye disease screening [58.861421804458395]
We propose a novel multi-modality evidential fusion pipeline for eye disease screening.
It provides a measure of confidence for each modality and elegantly integrates the multi-modality information.
Experimental results on both public and internal datasets demonstrate that our model excels in robustness.
arXiv Detail & Related papers (2024-05-28T13:27:30Z) - MOGAM: A Multimodal Object-oriented Graph Attention Model for Depression Detection [5.506046101113427]
We introduce a Multimodal Object-Oriented Graph Attention Model (MOGAM) for detecting depression in social media.
To ensure that our model can capture authentic symptoms of depression, we only include vlogs from users with a clinical diagnosis.
MOGAM achieved an accuracy of 0.871 and an F1-score of 0.888.
arXiv Detail & Related papers (2024-03-21T07:45:58Z) - Optimizing Skin Lesion Classification via Multimodal Data and Auxiliary
Task Integration [54.76511683427566]
This research introduces a novel multimodal method for classifying skin lesions, integrating smartphone-captured images with essential clinical and demographic information.
A distinctive aspect of this method is the integration of an auxiliary task focused on super-resolution image prediction.
The experimental evaluations have been conducted using the PAD-UFES20 dataset, applying various deep-learning architectures.
arXiv Detail & Related papers (2024-02-16T05:16:20Z) - Attention-Based Acoustic Feature Fusion Network for Depression Detection [11.972591489278988]
We present the Attention-Based Acoustic Feature Fusion Network (ABAFnet) for depression detection.
ABAFnet combines four different acoustic features into a comprehensive deep learning model, thereby effectively integrating and blending multi-tiered features.
We present a novel weight adjustment module for late fusion that boosts performance by efficaciously synthesizing these features.
arXiv Detail & Related papers (2023-08-24T00:31:51Z) - Depression Diagnosis and Analysis via Multimodal Multi-order Factor
Fusion [4.991507302519828]
Depression is a leading cause of death worldwide, and the diagnosis of depression is nontrivial.
We propose a multimodal multi-order factor fusion (MMFF) method for automatic diagnosis of depression.
arXiv Detail & Related papers (2022-12-31T17:13:06Z) - Multi-modal Depression Estimation based on Sub-attentional Fusion [29.74171323437029]
Failure to diagnose depression leads to over 280 million people suffering from this psychological disorder worldwide.
We tackle the task of automatically identifying depression from multi-modal data.
We introduce a sub-attention mechanism for linking heterogeneous information.
arXiv Detail & Related papers (2022-07-13T13:19:32Z) - Inheritance-guided Hierarchical Assignment for Clinical Automatic
Diagnosis [50.15205065710629]
Clinical diagnosis, which aims to assign diagnosis codes for a patient based on the clinical note, plays an essential role in clinical decision-making.
We propose a novel framework to combine the inheritance-guided hierarchical assignment and co-occurrence graph propagation for clinical automatic diagnosis.
arXiv Detail & Related papers (2021-01-27T13:16:51Z) - Deep Multi-task Learning for Depression Detection and Prediction in
Longitudinal Data [50.02223091927777]
Depression is among the most prevalent mental disorders, affecting millions of people of all ages globally.
Machine learning techniques have shown effective in enabling automated detection and prediction of depression for early intervention and treatment.
We introduce a novel deep multi-task recurrent neural network to tackle this challenge, in which depression classification is jointly optimized with two auxiliary tasks.
arXiv Detail & Related papers (2020-12-05T05:14:14Z) - Multimodal Depression Severity Prediction from medical bio-markers using
Machine Learning Tools and Technologies [0.0]
Depression has been a leading cause of mental-health illnesses across the world.
Using behavioural cues to automate depression diagnosis and stage prediction in recent years has relatively increased.
The absence of labelled behavioural datasets and a vast amount of possible variations prove to be a major challenge in accomplishing the task.
arXiv Detail & Related papers (2020-09-11T20:44:28Z) - Detecting Parkinsonian Tremor from IMU Data Collected In-The-Wild using
Deep Multiple-Instance Learning [59.74684475991192]
Parkinson's Disease (PD) is a slowly evolving neuro-logical disease that affects about 1% of the population above 60 years old.
PD symptoms include tremor, rigidity and braykinesia.
We present a method for automatically identifying tremorous episodes related to PD, based on IMU signals captured via a smartphone device.
arXiv Detail & Related papers (2020-05-06T09:02:30Z) - Multimodal Categorization of Crisis Events in Social Media [81.07061295887172]
We present a new multimodal fusion method that leverages both images and texts as input.
In particular, we introduce a cross-attention module that can filter uninformative and misleading components from weak modalities.
We show that our method outperforms the unimodal approaches and strong multimodal baselines by a large margin on three crisis-related tasks.
arXiv Detail & Related papers (2020-04-10T06:31:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.