Data Quality Matters: Suicide Intention Detection on Social Media Posts
Using a RoBERTa-CNN Model
- URL: http://arxiv.org/abs/2402.02262v1
- Date: Sat, 3 Feb 2024 20:58:09 GMT
- Title: Data Quality Matters: Suicide Intention Detection on Social Media Posts
Using a RoBERTa-CNN Model
- Authors: Emily Lin, Jian Sun, Hsingyu Chen, and Mohammad H. Mahoor
- Abstract summary: We present a novel approach to suicide detection using the cutting-edge RoBERTa-CNN model.
RoBERTa-CNN achieves 98% mean accuracy with the standard deviation (STD) of 0.0009.
It also reaches over 97.5% mean AUC value with an STD of 0.0013.
- Score: 39.143550443239064
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Suicide remains a global health concern for the field of health, which
urgently needs innovative approaches for early detection and intervention. In
this paper, we focus on identifying suicidal intentions in SuicideWatch Reddit
posts and present a novel approach to suicide detection using the cutting-edge
RoBERTa-CNN model, a variant of RoBERTa (Robustly optimized BERT approach).
RoBERTa is used for various Natural Language Processing (NLP) tasks, including
text classification and sentiment analysis. The effectiveness of the RoBERTa
lies in its ability to capture textual information and form semantic
relationships within texts. By adding the Convolution Neural Network (CNN)
layer to the original model, the RoBERTa enhances its ability to capture
important patterns from heavy datasets. To evaluate the RoBERTa-CNN, we
experimented on the Suicide and Depression Detection dataset and obtained solid
results. For example, RoBERTa-CNN achieves 98% mean accuracy with the standard
deviation (STD) of 0.0009. It also reaches over 97.5% mean AUC value with an
STD of 0.0013. In the meanwhile, RoBERTa-CNN outperforms competitive methods,
demonstrating the robustness and ability to capture nuanced linguistic patterns
for suicidal intentions. Therefore, RoBERTa-CNN can detect suicide intention on
text data very well.
Related papers
- Leveraging Large Language Models for Suicide Detection on Social Media with Limited Labels [3.1399304968349186]
This paper explores the use of Large Language Models (LLMs) to automatically detect suicidal content in text-based social media posts.
We develop an ensemble approach involving prompting with Qwen2-72B-Instruct, and using fine-tuned models such as Llama3-8B, Llama3.1-8B, and Gemma2-9B.
Experimental results show that the ensemble model significantly improves the detection accuracy, by 5% points compared with the individual models.
arXiv Detail & Related papers (2024-10-06T14:45:01Z) - Harnessing Neuron Stability to Improve DNN Verification [42.65507402735545]
We present VeriStable, a novel extension of recently proposed DPLL-based constraint DNN verification approach.
We evaluate the effectiveness of VeriStable across a range of challenging benchmarks including fully-connected feed networks (FNNs), convolutional neural networks (CNNs) and residual networks (ResNets)
Preliminary results show that VeriStable is competitive and outperforms state-of-the-art verification tools, including $alpha$-$beta$-CROWN and MN-BaB, the first and second performers of the VNN-COMP, respectively.
arXiv Detail & Related papers (2024-01-19T23:48:04Z) - GOODAT: Towards Test-time Graph Out-of-Distribution Detection [103.40396427724667]
Graph neural networks (GNNs) have found widespread application in modeling graph data across diverse domains.
Recent studies have explored graph OOD detection, often focusing on training a specific model or modifying the data on top of a well-trained GNN.
This paper introduces a data-centric, unsupervised, and plug-and-play solution that operates independently of training data and modifications of GNN architecture.
arXiv Detail & Related papers (2024-01-10T08:37:39Z) - Uncertainty Quantification over Graph with Conformalized Graph Neural
Networks [52.20904874696597]
Graph Neural Networks (GNNs) are powerful machine learning prediction models on graph-structured data.
GNNs lack rigorous uncertainty estimates, limiting their reliable deployment in settings where the cost of errors is significant.
We propose conformalized GNN (CF-GNN), extending conformal prediction (CP) to graph-based models for guaranteed uncertainty estimates.
arXiv Detail & Related papers (2023-05-23T21:38:23Z) - UPNet: Uncertainty-based Picking Deep Learning Network for Robust First Break Picking [6.380128763476294]
First break (FB) picking is a crucial aspect in the determination of subsurface velocity models.
Deep neural networks (DNNs) have been proposed to accelerate this processing.
We introduce uncertainty quantification into the FB picking task and propose a novel uncertainty-based deep learning network called UPNet.
arXiv Detail & Related papers (2023-05-23T08:13:09Z) - Energy-based Out-of-Distribution Detection for Graph Neural Networks [76.0242218180483]
We propose a simple, powerful and efficient OOD detection model for GNN-based learning on graphs, which we call GNNSafe.
GNNSafe achieves up to $17.0%$ AUROC improvement over state-of-the-arts and it could serve as simple yet strong baselines in such an under-developed area.
arXiv Detail & Related papers (2023-02-06T16:38:43Z) - Out-of-Distribution Detection with Hilbert-Schmidt Independence
Optimization [114.43504951058796]
Outlier detection tasks have been playing a critical role in AI safety.
Deep neural network classifiers usually tend to incorrectly classify out-of-distribution (OOD) inputs into in-distribution classes with high confidence.
We propose an alternative probabilistic paradigm that is both practically useful and theoretically viable for the OOD detection tasks.
arXiv Detail & Related papers (2022-09-26T15:59:55Z) - Labeling-Free Comparison Testing of Deep Learning Models [28.47632100019289]
We propose a labeling-free comparison testing approach to overcome the limitations of labeling effort and sampling randomness.
Our approach outperforms the baseline methods by up to 0.74 and 0.53 on Spearman's correlation and Kendall's $tau$, regardless of the dataset and distribution shift.
arXiv Detail & Related papers (2022-04-08T10:55:45Z) - A Fully Spiking Hybrid Neural Network for Energy-Efficient Object
Detection [6.792495874038191]
Fully Spiking Hybrid Neural Network (FSHNN) for energy-efficient and robust object detection.
Network architecture is based on Convolutional SNN using leaky-integrate-fire neuron models.
arXiv Detail & Related papers (2021-04-21T18:39:32Z) - Deep Time Delay Neural Network for Speech Enhancement with Full Data
Learning [60.20150317299749]
This paper proposes a deep time delay neural network (TDNN) for speech enhancement with full data learning.
To make full use of the training data, we propose a full data learning method for speech enhancement.
arXiv Detail & Related papers (2020-11-11T06:32:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.