Large Language Models for Depression Recognition in Spoken Language Integrating Psychological Knowledge
- URL: http://arxiv.org/abs/2505.22863v1
- Date: Wed, 28 May 2025 20:53:05 GMT
- Title: Large Language Models for Depression Recognition in Spoken Language Integrating Psychological Knowledge
- Authors: Yupei Li, Shuaijie Shao, Manuel Milling, Björn W. Schuller,
- Abstract summary: Depression is a growing concern gaining attention in both public discourse and AI research.<n>We present to the best of our knowledge, the first application of large language models to multimodal depression detection.<n>We propose a novel strategy for incorporating psychological knowledge into LLMs to enhance diagnostic performance.
- Score: 38.52606718906513
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Depression is a growing concern gaining attention in both public discourse and AI research. While deep neural networks (DNNs) have been used for recognition, they still lack real-world effectiveness. Large language models (LLMs) show strong potential but require domain-specific fine-tuning and struggle with non-textual cues. Since depression is often expressed through vocal tone and behaviour rather than explicit text, relying on language alone is insufficient. Diagnostic accuracy also suffers without incorporating psychological expertise. To address these limitations, we present, to the best of our knowledge, the first application of LLMs to multimodal depression detection using the DAIC-WOZ dataset. We extract the audio features using the pre-trained model Wav2Vec, and mapped it to text-based LLMs for further processing. We also propose a novel strategy for incorporating psychological knowledge into LLMs to enhance diagnostic performance, specifically using a question and answer set to grant authorised knowledge to LLMs. Our approach yields a notable improvement in both Mean Absolute Error (MAE) and Root Mean Square Error (RMSE) compared to a base score proposed by the related original paper. The codes are available at https://github.com/myxp-lyp/Depression-detection.git
Related papers
- MLlm-DR: Towards Explainable Depression Recognition with MultiModal Large Language Models [28.873959594226605]
Automated depression diagnosis aims to analyze multimodal information from interview videos to predict participants' depression scores.<n>Previous studies often lack clear explanations of how these scores were determined, limiting their adoption in clinical practice.<n>We propose a novel multimodal large language model (MLlm-DR) that can understand multimodal information inputs and supports explainable depression diagnosis.
arXiv Detail & Related papers (2025-07-08T01:56:39Z) - Explainable Depression Detection in Clinical Interviews with Personalized Retrieval-Augmented Generation [32.163466666512996]
Depression is a widespread mental health disorder, and clinical interviews are the gold standard for assessment.<n>Current systems mainly employ black-box neural networks, which lack interpretability.<n>We propose RED, a Retrieval-augmented generation framework for Explainable depression Detection.
arXiv Detail & Related papers (2025-03-03T08:59:34Z) - LlaMADRS: Prompting Large Language Models for Interview-Based Depression Assessment [75.44934940580112]
This study introduces LlaMADRS, a novel framework leveraging open-source Large Language Models (LLMs) to automate depression severity assessment.<n>We employ a zero-shot prompting strategy with carefully designed cues to guide the model in interpreting and scoring transcribed clinical interviews.<n>Our approach, tested on 236 real-world interviews, demonstrates strong correlations with clinician assessments.
arXiv Detail & Related papers (2025-01-07T08:49:04Z) - A BERT-Based Summarization approach for depression detection [1.7363112470483526]
Depression is a globally prevalent mental disorder with potentially severe repercussions if not addressed.
Machine learning and artificial intelligence can autonomously detect depression indicators from diverse data sources.
Our study proposes text summarization as a preprocessing technique to diminish the length and intricacies of input texts.
arXiv Detail & Related papers (2024-09-13T02:14:34Z) - Negation Blindness in Large Language Models: Unveiling the NO Syndrome in Image Generation [63.064204206220936]
Foundational Large Language Models (LLMs) have changed the way we perceive technology.
They have been shown to excel in tasks ranging from poem writing to coding to essay generation and puzzle solving.
With the incorporation of image generation capability, they have become more comprehensive and versatile AI tools.
Currently identified flaws include hallucination, biases, and bypassing restricted commands to generate harmful content.
arXiv Detail & Related papers (2024-08-27T14:40:16Z) - Knowledge Graph-Enhanced Large Language Models via Path Selection [58.228392005755026]
Large Language Models (LLMs) have shown unprecedented performance in various real-world applications.
LLMs are known to generate factually inaccurate outputs, a.k.a. the hallucination problem.
We propose a principled framework KELP with three stages to handle the above problems.
arXiv Detail & Related papers (2024-06-19T21:45:20Z) - Depression Detection on Social Media with Large Language Models [23.075317886505193]
Depression detection aims to determine whether an individual suffers from depression by analyzing their history of posts on social media.
We propose a novel depression detection system called DORIS, combining medical knowledge and the recent advances in large language models.
arXiv Detail & Related papers (2024-03-16T01:01:16Z) - Tuning-Free Accountable Intervention for LLM Deployment -- A
Metacognitive Approach [55.613461060997004]
Large Language Models (LLMs) have catalyzed transformative advances across a spectrum of natural language processing tasks.
We propose an innovative textitmetacognitive approach, dubbed textbfCLEAR, to equip LLMs with capabilities for self-aware error identification and correction.
arXiv Detail & Related papers (2024-03-08T19:18:53Z) - When LLMs Meets Acoustic Landmarks: An Efficient Approach to Integrate Speech into Large Language Models for Depression Detection [17.018248242646365]
Depression is a critical concern in global mental health, prompting extensive research into AI-based detection methods.
Large Language Models (LLMs) stand out for their versatility in mental healthcare applications.
We present an innovative approach to integrating acoustic speech information into the LLMs framework for multimodal depression detection.
arXiv Detail & Related papers (2024-02-17T09:39:46Z) - Enhancing Uncertainty-Based Hallucination Detection with Stronger Focus [99.33091772494751]
Large Language Models (LLMs) have gained significant popularity for their impressive performance across diverse fields.
LLMs are prone to hallucinate untruthful or nonsensical outputs that fail to meet user expectations.
We propose a novel reference-free, uncertainty-based method for detecting hallucinations in LLMs.
arXiv Detail & Related papers (2023-11-22T08:39:17Z) - Are Large Language Models Really Robust to Word-Level Perturbations? [68.60618778027694]
We propose a novel rational evaluation approach that leverages pre-trained reward models as diagnostic tools.
Longer conversations manifest the comprehensive grasp of language models in terms of their proficiency in understanding questions.
Our results demonstrate that LLMs frequently exhibit vulnerability to word-level perturbations that are commonplace in daily language usage.
arXiv Detail & Related papers (2023-09-20T09:23:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.