Boosting Distress Support Dialogue Responses with Motivational
Interviewing Strategy
- URL: http://arxiv.org/abs/2305.10195v1
- Date: Wed, 17 May 2023 13:18:28 GMT
- Title: Boosting Distress Support Dialogue Responses with Motivational
Interviewing Strategy
- Authors: Anuradha Welivita and Pearl Pu
- Abstract summary: We show how some response types could be rephrased into a more MI adherent form.
We build several rephrasers by fine-tuning Blender and GPT3 to rephrase MI non-adherent "Advise without permission" responses into "Advise with permission"
- Score: 4.264192013842096
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: AI-driven chatbots have become an emerging solution to address psychological
distress. Due to the lack of psychotherapeutic data, researchers use dialogues
scraped from online peer support forums to train them. But since the responses
in such platforms are not given by professionals, they contain both conforming
and non-conforming responses. In this work, we attempt to recognize these
conforming and non-conforming response types present in online distress-support
dialogues using labels adapted from a well-established behavioral coding scheme
named Motivational Interviewing Treatment Integrity (MITI) code and show how
some response types could be rephrased into a more MI adherent form that can,
in turn, enable chatbot responses to be more compliant with the MI strategy. As
a proof of concept, we build several rephrasers by fine-tuning Blender and GPT3
to rephrase MI non-adherent "Advise without permission" responses into "Advise
with permission". We show how this can be achieved with the construction of
pseudo-parallel corpora avoiding costs for human labor. Through automatic and
human evaluation we show that in the presence of less training data, techniques
such as prompting and data augmentation can be used to produce substantially
good rephrasings that reflect the intended style and preserve the content of
the original text.
Related papers
- Prompt Engineering a Schizophrenia Chatbot: Utilizing a Multi-Agent Approach for Enhanced Compliance with Prompt Instructions [0.0699049312989311]
Patients with schizophrenia often present with cognitive impairments that may hinder their ability to learn about their condition.
While Large Language Models (LLMs) have the potential to make topical mental health information more accessible and engaging, their black-box nature raises concerns about ethics and safety.
arXiv Detail & Related papers (2024-10-10T09:49:24Z) - Towards Empathetic Conversational Recommender Systems [77.53167131692]
We propose an empathetic conversational recommender (ECR) framework.
ECR contains two main modules: emotion-aware item recommendation and emotion-aligned response generation.
Our experiments on the ReDial dataset validate the efficacy of our framework in enhancing recommendation accuracy and improving user satisfaction.
arXiv Detail & Related papers (2024-08-30T15:43:07Z) - Mitigating Unhelpfulness in Emotional Support Conversations with Multifaceted AI Feedback [9.57004333812654]
We introduce a novel model-agnostic framework named mitigating unhelpfulness with multifaceted AI feedback for emotional support (Muffin)
Muffin employs a multifaceted AI feedback module to assess the helpfulness of responses generated by a specific model with consideration of multiple factors.
Results demonstrate that Muffin effectively mitigates the generation of unhelpful responses while slightly increasing response fluency and relevance.
arXiv Detail & Related papers (2024-01-11T14:07:47Z) - AutoReply: Detecting Nonsense in Dialogue Introspectively with
Discriminative Replies [71.62832112141913]
We show that dialogue models can detect errors in their own messages introspectively, by calculating the likelihood of replies that are indicative of poor messages.
We first show that hand-crafted replies can be effective for the task of detecting nonsense in applications as complex as Diplomacy.
We find that AutoReply-generated replies outperform handcrafted replies and perform on par with carefully fine-tuned large supervised models.
arXiv Detail & Related papers (2022-11-22T22:31:34Z) - A Systematic Evaluation of Response Selection for Open Domain Dialogue [36.88551817451512]
We curated a dataset where responses from multiple response generators produced for the same dialog context are manually annotated as appropriate (positive) and inappropriate (negative)
We conduct a systematic evaluation of state-of-the-art methods for response selection, and demonstrate that both strategies of using multiple positive candidates and using manually verified hard negative candidates can bring in significant performance improvement in comparison to using the adversarial training data, e.g., increase of 3% and 13% in Recall@1 score, respectively.
arXiv Detail & Related papers (2022-08-08T19:33:30Z) - What is wrong with you?: Leveraging User Sentiment for Automatic Dialog
Evaluation [73.03318027164605]
We propose to use information that can be automatically extracted from the next user utterance as a proxy to measure the quality of the previous system response.
Our model generalizes across both spoken and written open-domain dialog corpora collected from real and paid users.
arXiv Detail & Related papers (2022-03-25T22:09:52Z) - MISC: A MIxed Strategy-Aware Model Integrating COMET for Emotional
Support Conversation [64.37111498077866]
We propose a novel model for emotional support conversation.
It infers the user's fine-grained emotional status, and then responds skillfully using a mixture of strategy.
Experimental results on the benchmark dataset demonstrate the effectiveness of our method.
arXiv Detail & Related papers (2022-03-25T10:32:04Z) - Unsupervised Contextual Paraphrase Generation using Lexical Control and
Reinforcement Learning [3.2811284938530636]
We propose an unsupervised frame-work to generate contextual paraphrases using autoregressive models.
We also propose an automated metric based on Semantic Similarity, Textual Entailment, Expression Diversity and Fluency to evaluate the quality of contextual paraphrases.
arXiv Detail & Related papers (2021-03-23T18:22:03Z) - Saying No is An Art: Contextualized Fallback Responses for Unanswerable
Dialogue Queries [3.593955557310285]
Most dialogue systems rely on hybrid approaches for generating a set of ranked responses.
We design a neural approach which generates responses which are contextually aware with the user query.
Our simple approach makes use of rules over dependency parses and a text-to-text transformer fine-tuned on synthetic data of question-response pairs.
arXiv Detail & Related papers (2020-12-03T12:34:22Z) - Generating Dialogue Responses from a Semantic Latent Space [75.18449428414736]
We propose an alternative to the end-to-end classification on vocabulary.
We learn the pair relationship between the prompts and responses as a regression task on a latent space.
Human evaluation showed that learning the task on a continuous space can generate responses that are both relevant and informative.
arXiv Detail & Related papers (2020-10-04T19:06:16Z) - Learning an Unreferenced Metric for Online Dialogue Evaluation [53.38078951628143]
We propose an unreferenced automated evaluation metric that uses large pre-trained language models to extract latent representations of utterances.
We show that our model achieves higher correlation with human annotations in an online setting, while not requiring true responses for comparison during inference.
arXiv Detail & Related papers (2020-05-01T20:01:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.