Utilizing Speech Emotion Recognition and Recommender Systems for
Negative Emotion Handling in Therapy Chatbots
- URL: http://arxiv.org/abs/2311.11116v1
- Date: Sat, 18 Nov 2023 16:35:55 GMT
- Title: Utilizing Speech Emotion Recognition and Recommender Systems for
Negative Emotion Handling in Therapy Chatbots
- Authors: Farideh Majidi, Marzieh Bahrami
- Abstract summary: This paper proposes an approach to enhance therapy chatbots with auditory perception, enabling them to understand users' feelings and provide human-like empathy.
The proposed method incorporates speech emotion recognition (SER) techniques using CNN models and the ShEMO dataset.
To provide a more immersive and empathetic user experience, a text-to-speech model called GlowTTS is integrated.
- Score: 0.0
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Emotional well-being significantly influences mental health and overall
quality of life. As therapy chatbots become increasingly prevalent, their
ability to comprehend and respond empathetically to users' emotions remains
limited. This paper addresses this limitation by proposing an approach to
enhance therapy chatbots with auditory perception, enabling them to understand
users' feelings and provide human-like empathy. The proposed method
incorporates speech emotion recognition (SER) techniques using Convolutional
Neural Network (CNN) models and the ShEMO dataset to accurately detect and
classify negative emotions, including anger, fear, and sadness. The SER model
achieves a validation accuracy of 88%, demonstrating its effectiveness in
recognizing emotional states from speech signals. Furthermore, a recommender
system is developed, leveraging the SER model's output to generate personalized
recommendations for managing negative emotions, for which a new bilingual
dataset was generated as well since there is no such dataset available for this
task. The recommender model achieves an accuracy of 98% by employing a
combination of global vectors for word representation (GloVe) and LSTM models.
To provide a more immersive and empathetic user experience, a text-to-speech
model called GlowTTS is integrated, enabling the therapy chatbot to audibly
communicate the generated recommendations to users in both English and Persian.
The proposed approach offers promising potential to enhance therapy chatbots by
providing them with the ability to recognize and respond to users' emotions,
ultimately improving the delivery of mental health support for both English and
Persian-speaking users.
Related papers
- Emotion-Aware Response Generation Using Affect-Enriched Embeddings with LLMs [0.585143166250719]
This study addresses the challenge of enhancing the emotional and contextual understanding of large language models (LLMs) in psychiatric applications.
We introduce a novel framework that integrates multiple emotion lexicons, with state-of-the-art LLMs such as LLAMA 2, Flan-T5, ChatGPT 3.0, and ChatGPT 4.0.
The primary dataset comprises over 2,000 therapy session transcripts from the Counseling and Psychotherapy database, covering discussions on anxiety, depression, trauma, and addiction.
arXiv Detail & Related papers (2024-10-02T08:01:05Z) - Towards Empathetic Conversational Recommender Systems [77.53167131692]
We propose an empathetic conversational recommender (ECR) framework.
ECR contains two main modules: emotion-aware item recommendation and emotion-aligned response generation.
Our experiments on the ReDial dataset validate the efficacy of our framework in enhancing recommendation accuracy and improving user satisfaction.
arXiv Detail & Related papers (2024-08-30T15:43:07Z) - Personality-affected Emotion Generation in Dialog Systems [67.40609683389947]
We propose a new task, Personality-affected Emotion Generation, to generate emotion based on the personality given to the dialog system.
We analyze the challenges in this task, i.e., (1) heterogeneously integrating personality and emotional factors and (2) extracting multi-granularity emotional information in the dialog context.
Results suggest that by adopting our method, the emotion generation performance is improved by 13% in macro-F1 and 5% in weighted-F1 from the BERT-base model.
arXiv Detail & Related papers (2024-04-03T08:48:50Z) - HealMe: Harnessing Cognitive Reframing in Large Language Models for Psychotherapy [25.908522131646258]
We unveil the Helping and Empowering through Adaptive Language in Mental Enhancement (HealMe) model.
This novel cognitive reframing therapy method effectively addresses deep-rooted negative thoughts and fosters rational, balanced perspectives.
We adopt the first comprehensive and expertly crafted psychological evaluation metrics, specifically designed to rigorously assess the performance of cognitive reframing.
arXiv Detail & Related papers (2024-02-26T09:10:34Z) - Towards Multi-Turn Empathetic Dialogs with Positive Emotion Elicitation [39.747587984500406]
This paper presents a novel task of empathetic dialog generation with positive emotion elicitation.
The agent conducts empathetic responses along with the target of eliciting the user's positive emotions in the multi-turn dialog.
We collect a large-scale emotional dialog dataset with positive emotion elicitation, called PosEmoDial.
arXiv Detail & Related papers (2022-04-22T05:32:08Z) - MISC: A MIxed Strategy-Aware Model Integrating COMET for Emotional
Support Conversation [64.37111498077866]
We propose a novel model for emotional support conversation.
It infers the user's fine-grained emotional status, and then responds skillfully using a mixture of strategy.
Experimental results on the benchmark dataset demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2022-03-25T10:32:04Z) - Multimodal Emotion Recognition using Transfer Learning from Speaker
Recognition and BERT-based models [53.31917090073727]
We propose a neural network-based emotion recognition framework that uses a late fusion of transfer-learned and fine-tuned models from speech and text modalities.
We evaluate the effectiveness of our proposed multimodal approach on the interactive emotional dyadic motion capture dataset.
arXiv Detail & Related papers (2022-02-16T00:23:42Z) - EMOVIE: A Mandarin Emotion Speech Dataset with a Simple Emotional
Text-to-Speech Model [56.75775793011719]
We introduce and publicly release a Mandarin emotion speech dataset including 9,724 samples with audio files and its emotion human-labeled annotation.
Unlike those models which need additional reference audio as input, our model could predict emotion labels just from the input text and generate more expressive speech conditioned on the emotion embedding.
In the experiment phase, we first validate the effectiveness of our dataset by an emotion classification task. Then we train our model on the proposed dataset and conduct a series of subjective evaluations.
arXiv Detail & Related papers (2021-06-17T08:34:21Z) - Emotion-aware Chat Machine: Automatic Emotional Response Generation for
Human-like Emotional Interaction [55.47134146639492]
This article proposes a unifed end-to-end neural architecture, which is capable of simultaneously encoding the semantics and the emotions in a post.
Experiments on real-world data demonstrate that the proposed method outperforms the state-of-the-art methods in terms of both content coherence and emotion appropriateness.
arXiv Detail & Related papers (2021-06-06T06:26:15Z) - Speech-Based Emotion Recognition using Neural Networks and Information
Visualization [1.52292571922932]
We propose a tool which enables users to take speech samples and identify a range of emotions from audio elements.
The dashboard is designed based on local therapists' needs for intuitive representations of speech data.
arXiv Detail & Related papers (2020-10-28T20:57:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.