Harnessing Large Language Models Over Transformer Models for Detecting
Bengali Depressive Social Media Text: A Comprehensive Study
- URL: http://arxiv.org/abs/2401.07310v1
- Date: Sun, 14 Jan 2024 15:15:58 GMT
- Title: Harnessing Large Language Models Over Transformer Models for Detecting
Bengali Depressive Social Media Text: A Comprehensive Study
- Authors: Ahmadul Karim Chowdhury, Md. Saidur Rahman Sujon, Md. Shirajus Salekin
Shafi, Tasin Ahmmad, Sifat Ahmed, Khan Md Hasib, Faisal Muhammad Shah
- Abstract summary: This work focuses on early detection of depression using LLMs such as GPT 3.5, GPT 4 and our proposed GPT 3.5 fine-tuned model DepGPT.
The study categorized Reddit and X datasets into "Depressive" and "Non-Depressive" segments, translated into Bengali by native speakers with expertise in mental health.
Our work provides full architecture details for each model and a methodical way to assess their performance in Bengali depressive text categorization.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: In an era where the silent struggle of underdiagnosed depression pervades
globally, our research delves into the crucial link between mental health and
social media. This work focuses on early detection of depression, particularly
in extroverted social media users, using LLMs such as GPT 3.5, GPT 4 and our
proposed GPT 3.5 fine-tuned model DepGPT, as well as advanced Deep learning
models(LSTM, Bi-LSTM, GRU, BiGRU) and Transformer models(BERT, BanglaBERT,
SahajBERT, BanglaBERT-Base). The study categorized Reddit and X datasets into
"Depressive" and "Non-Depressive" segments, translated into Bengali by native
speakers with expertise in mental health, resulting in the creation of the
Bengali Social Media Depressive Dataset (BSMDD). Our work provides full
architecture details for each model and a methodical way to assess their
performance in Bengali depressive text categorization using zero-shot and
few-shot learning techniques. Our work demonstrates the superiority of
SahajBERT and Bi-LSTM with FastText embeddings in their respective domains also
tackles explainability issues with transformer models and emphasizes the
effectiveness of LLMs, especially DepGPT, demonstrating flexibility and
competence in a range of learning contexts. According to the experiment
results, the proposed model, DepGPT, outperformed not only Alpaca Lora 7B in
zero-shot and few-shot scenarios but also every other model, achieving a
near-perfect accuracy of 0.9796 and an F1-score of 0.9804, high recall, and
exceptional precision. Although competitive, GPT-3.5 Turbo and Alpaca Lora 7B
show relatively poorer effectiveness in zero-shot and few-shot situations. The
work emphasizes the effectiveness and flexibility of LLMs in a variety of
linguistic circumstances, providing insightful information about the complex
field of depression detection models.
Related papers
- A BERT-Based Summarization approach for depression detection [1.7363112470483526]
Depression is a globally prevalent mental disorder with potentially severe repercussions if not addressed.
Machine learning and artificial intelligence can autonomously detect depression indicators from diverse data sources.
Our study proposes text summarization as a preprocessing technique to diminish the length and intricacies of input texts.
arXiv Detail & Related papers (2024-09-13T02:14:34Z) - Enhancing Depressive Post Detection in Bangla: A Comparative Study of TF-IDF, BERT and FastText Embeddings [0.0]
This study introduces a well-grounded approach to identify depressive social media posts in Bangla.
The dataset used in this work, annotated by domain experts, includes both depressive and non-depressive posts.
To address the issue of class imbalance, we utilised random oversampling for the minority class.
arXiv Detail & Related papers (2024-07-12T11:40:17Z) - How Easy is It to Fool Your Multimodal LLMs? An Empirical Analysis on Deceptive Prompts [54.07541591018305]
We present MAD-Bench, a benchmark that contains 1000 test samples divided into 5 categories, such as non-existent objects, count of objects, and spatial relationship.
We provide a comprehensive analysis of popular MLLMs, ranging from GPT-4v, Reka, Gemini-Pro, to open-sourced models, such as LLaVA-NeXT and MiniCPM-Llama3.
While GPT-4o achieves 82.82% accuracy on MAD-Bench, the accuracy of any other model in our experiments ranges from 9% to 50%.
arXiv Detail & Related papers (2024-02-20T18:31:27Z) - Calibration of Transformer-based Models for Identifying Stress and
Depression in Social Media [0.0]
We present the first study in the task of depression and stress detection in social media, which injects extra linguistic information in transformer-based models.
Specifically, the proposed approach employs a Multimodal Adaptation Gate for creating the combined embeddings, which are given as input to a BERT (or MentalBERT) model.
We test our proposed approaches in three publicly available datasets and demonstrate that the integration of linguistic features into transformer-based models presents a surge in the performance.
arXiv Detail & Related papers (2023-05-26T10:19:04Z) - Evaluating Psychological Safety of Large Language Models [72.88260608425949]
We designed unbiased prompts to evaluate the psychological safety of large language models (LLMs)
We tested five different LLMs by using two personality tests: Short Dark Triad (SD-3) and Big Five Inventory (BFI)
Despite being instruction fine-tuned with safety metrics to reduce toxicity, InstructGPT, GPT-3.5, and GPT-4 still showed dark personality patterns.
Fine-tuning Llama-2-chat-7B with responses from BFI using direct preference optimization could effectively reduce the psychological toxicity of the model.
arXiv Detail & Related papers (2022-12-20T18:45:07Z) - Prompting GPT-3 To Be Reliable [117.23966502293796]
This work decomposes reliability into four facets: generalizability, fairness, calibration, and factuality.
We find that GPT-3 outperforms smaller-scale supervised models by large margins on all these facets.
arXiv Detail & Related papers (2022-10-17T14:52:39Z) - DECK: Behavioral Tests to Improve Interpretability and Generalizability
of BERT Models Detecting Depression from Text [4.269268432906194]
Models that accurately detect depression from text are important tools for addressing the post-pandemic mental health crisis.
BERT-based classifiers' promising performance and the off-the-shelf availability make them great candidates for this task.
We introduce the DECK (DEpression ChecKlist), depression-specific model behavioural tests that allow better interpretability.
arXiv Detail & Related papers (2022-09-12T14:39:46Z) - Prediction of Depression Severity Based on the Prosodic and Semantic
Features with Bidirectional LSTM and Time Distributed CNN [14.994852548758825]
We propose an attention-based multimodality speech and text representation for depression prediction.
Our model is trained to estimate the depression severity of participants using the Distress Analysis Interview Corpus-Wizard of Oz dataset.
Experiments show statistically significant improvements over previous works.
arXiv Detail & Related papers (2022-02-25T01:42:29Z) - Fine-Tuning Large Neural Language Models for Biomedical Natural Language
Processing [55.52858954615655]
We conduct a systematic study on fine-tuning stability in biomedical NLP.
We show that finetuning performance may be sensitive to pretraining settings, especially in low-resource domains.
We show that these techniques can substantially improve fine-tuning performance for lowresource biomedical NLP applications.
arXiv Detail & Related papers (2021-12-15T04:20:35Z) - What do Compressed Large Language Models Forget? Robustness Challenges
in Model Compression [68.82486784654817]
We study two popular model compression techniques including knowledge distillation and pruning.
We show that compressed models are significantly less robust than their PLM counterparts on adversarial test sets.
We develop a regularization strategy for model compression based on sample uncertainty.
arXiv Detail & Related papers (2021-10-16T00:20:04Z) - Joint Energy-based Model Training for Better Calibrated Natural Language
Understanding Models [61.768082640087]
We explore joint energy-based model (EBM) training during the finetuning of pretrained text encoders for natural language understanding tasks.
Experiments show that EBM training can help the model reach a better calibration that is competitive to strong baselines.
arXiv Detail & Related papers (2021-01-18T01:41:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.