Explainable Depression Detection with Multi-Modalities Using a Hybrid
Deep Learning Model on Social Media
- URL: http://arxiv.org/abs/2007.02847v2
- Date: Wed, 28 Apr 2021 09:33:07 GMT
- Title: Explainable Depression Detection with Multi-Modalities Using a Hybrid
Deep Learning Model on Social Media
- Authors: Hamad Zogan, Imran Razzak, Xianzhi Wang, Shoaib Jameel, Guandong Xu
- Abstract summary: We propose interpretive Multi-Modal Depression Detection with Hierarchical Attention Network MDHAN.
Our model helps improve predictive performance when detecting depression in users who are posting messages publicly on social media.
- Score: 21.619614611039257
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Model interpretability has become important to engenders appropriate user
trust by providing the insight into the model prediction. However, most of the
existing machine learning methods provide no interpretability for depression
prediction, hence their predictions are obscure to human. In this work, we
propose interpretive Multi-Modal Depression Detection with Hierarchical
Attention Network MDHAN, for detection depressed users on social media and
explain the model prediction. We have considered user posts along with
Twitter-based multi-modal features, specifically, we encode user posts using
two levels of attention mechanisms applied at the tweet-level and word-level,
calculate each tweet and words' importance, and capture semantic sequence
features from the user timelines (posts). Our experiments show that MDHAN
outperforms several popular and robust baseline methods, demonstrating the
effectiveness of combining deep learning with multi-modal features. We also
show that our model helps improve predictive performance when detecting
depression in users who are posting messages publicly on social media. MDHAN
achieves excellent performance and ensures adequate evidence to explain the
prediction.
Related papers
- Sparse Prototype Network for Explainable Pedestrian Behavior Prediction [60.80524827122901]
We present Sparse Prototype Network (SPN), an explainable method designed to simultaneously predict a pedestrian's future action, trajectory, and pose.
Regularized by mono-semanticity and clustering constraints, the prototypes learn consistent and human-understandable features.
arXiv Detail & Related papers (2024-10-16T03:33:40Z) - Beyond Text: Leveraging Multi-Task Learning and Cognitive Appraisal Theory for Post-Purchase Intention Analysis [10.014248704653]
This study evaluates multi-task learning frameworks grounded in Cognitive Appraisal Theory to predict user behavior.
Our experiments show that users' language and traits improve predictions above and beyond models predicting only from text.
arXiv Detail & Related papers (2024-07-11T04:57:52Z) - Decoding the Silent Majority: Inducing Belief Augmented Social Graph
with Large Language Model for Response Forecasting [74.68371461260946]
SocialSense is a framework that induces a belief-centered graph on top of an existent social network, along with graph-based propagation to capture social dynamics.
Our method surpasses existing state-of-the-art in experimental evaluations for both zero-shot and supervised settings.
arXiv Detail & Related papers (2023-10-20T06:17:02Z) - Detecting and Grounding Multi-Modal Media Manipulation and Beyond [93.08116982163804]
We highlight a new research problem for multi-modal fake media, namely Detecting and Grounding Multi-Modal Media Manipulation (DGM4)
DGM4 aims to not only detect the authenticity of multi-modal media, but also ground the manipulated content.
We propose a novel HierArchical Multi-modal Manipulation rEasoning tRansformer (HAMMER) to fully capture the fine-grained interaction between different modalities.
arXiv Detail & Related papers (2023-09-25T15:05:46Z) - Depression detection in social media posts using affective and social
norm features [84.12658971655253]
We propose a deep architecture for depression detection from social media posts.
We incorporate profanity and morality features of posts and words in our architecture using a late fusion scheme.
The inclusion of the proposed features yields state-of-the-art results in both settings.
arXiv Detail & Related papers (2023-03-24T21:26:27Z) - It's Just a Matter of Time: Detecting Depression with Time-Enriched
Multimodal Transformers [24.776445591293186]
We propose a flexible time-enriched multimodal transformer architecture for detecting depression from social media posts.
Our model operates directly at the user-level, and we enrich it with the relative time between posts by using time2vec positional embeddings.
We show that our method, using EmoBERTa and CLIP embeddings, surpasses other methods on two multimodal datasets.
arXiv Detail & Related papers (2023-01-13T09:40:19Z) - Perceptual Score: What Data Modalities Does Your Model Perceive? [73.75255606437808]
We introduce the perceptual score, a metric that assesses the degree to which a model relies on the different subsets of the input features.
We find that recent, more accurate multi-modal models for visual question-answering tend to perceive the visual data less than their predecessors.
Using the perceptual score also helps to analyze model biases by decomposing the score into data subset contributions.
arXiv Detail & Related papers (2021-10-27T12:19:56Z) - A Framework for Multisensory Foresight for Embodied Agents [11.351546861334292]
Predicting future sensory states is crucial for learning agents such as robots, drones, and autonomous vehicles.
In this paper, we couple multiple sensory modalities with exploratory actions and propose a predictive neural network architecture to address this problem.
The framework was tested and validated with a dataset containing 4 sensory modalities (vision, haptic, audio, and tactile) on a humanoid robot performing 9 behaviors multiple times on a large set of objects.
arXiv Detail & Related papers (2021-09-15T20:20:04Z) - Interpretable Social Anchors for Human Trajectory Forecasting in Crowds [84.20437268671733]
We propose a neural network-based system to predict human trajectory in crowds.
We learn interpretable rule-based intents, and then utilise the expressibility of neural networks to model scene-specific residual.
Our architecture is tested on the interaction-centric benchmark TrajNet++.
arXiv Detail & Related papers (2021-05-07T09:22:34Z) - A Novel Sentiment Analysis Engine for Preliminary Depression Status
Estimation on Social Media [0.0]
We propose a cloud-based smartphone application, with a deep learning-based backend to primarily perform depression detection on Twitter social media.
A psychologist could leverage the application to assess the patient's depression status prior to counseling, which provides better insight into the mental health status of a patient.
The model achieved pinnacle results, with a testing accuracy of 87.23% and an AUC of 0.8621.
arXiv Detail & Related papers (2020-11-29T04:42:53Z) - Multimodal Depression Severity Prediction from medical bio-markers using
Machine Learning Tools and Technologies [0.0]
Depression has been a leading cause of mental-health illnesses across the world.
Using behavioural cues to automate depression diagnosis and stage prediction in recent years has relatively increased.
The absence of labelled behavioural datasets and a vast amount of possible variations prove to be a major challenge in accomplishing the task.
arXiv Detail & Related papers (2020-09-11T20:44:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.