Generative Adversarial Imitation Learning for Empathy-based AI
- URL: http://arxiv.org/abs/2105.13328v1
- Date: Thu, 27 May 2021 17:37:37 GMT
- Title: Generative Adversarial Imitation Learning for Empathy-based AI
- Authors: Pratyush Muthukumar, Karishma Muthukumar, Deepan Muthirayan, Pramod
Khargonekar
- Abstract summary: Generative adversarial imitation learning (GAIL) is a model-free algorithm that has been shown to provide strong results in imitating complex behaviors in high-dimensional environments.
In this paper, we utilize the GAIL model for text generation to develop empathy-based context-aware conversational AI.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Generative adversarial imitation learning (GAIL) is a model-free algorithm
that has been shown to provide strong results in imitating complex behaviors in
high-dimensional environments. In this paper, we utilize the GAIL model for
text generation to develop empathy-based context-aware conversational AI. Our
model uses an expert trajectory of empathetic prompt-response dialogues which
can accurately exhibit the correct empathetic emotion when generating a
response. The Generator of the GAIL model uses the GPT-2 sequential pre-trained
language model trained on 117 million parameters from 40 GB of internet data.
We propose a novel application of an approach used in transfer learning to fine
tune the GPT-2 model in order to generate concise, user-specific empathetic
responses validated against the Discriminator. Our novel GAIL model utilizes a
sentiment analysis history-based reinforcement learning approach to
empathetically respond to human interactions in a personalized manner. We find
that our model's response scores on various human-generated prompts collected
from the Facebook Empathetic Dialogues dataset outperform baseline
counterparts. Moreover, our model improves upon various history-based
conversational AI models developed recently, as our model's performance over a
sustained conversation of 3 or more interactions outperform similar
conversational AI models.
Related papers
- Promoting Open-domain Dialogue Generation through Learning Pattern
Information between Contexts and Responses [5.936682548344234]
This paper improves the quality of generated responses by learning the implicit pattern information between contexts and responses in the training samples.
We also design a response-aware mechanism for mining the implicit pattern information between contexts and responses so that the generated replies are more diverse and approximate to human replies.
arXiv Detail & Related papers (2023-09-06T08:11:39Z) - Adversarial Conversational Shaping for Intelligent Agents [0.0]
We study the performance of two models able to enhance an intelligent conversational agent through adversarial conversational shaping.
One model is able to assign rewards to both partially and fully generated text sequences.
We discuss performance with different training details : seq2seq [ 36] and transformers [37] in a reinforcement learning framework.
arXiv Detail & Related papers (2023-07-20T12:44:47Z) - BatGPT: A Bidirectional Autoregessive Talker from Generative Pre-trained
Transformer [77.28871523946418]
BatGPT is a large-scale language model designed and trained jointly by Wuhan University and Shanghai Jiao Tong University.
It is capable of generating highly natural and fluent text in response to various types of input, including text prompts, images, and audio.
arXiv Detail & Related papers (2023-07-01T15:10:01Z) - Robust Preference Learning for Storytelling via Contrastive
Reinforcement Learning [53.92465205531759]
Controlled automated story generation seeks to generate natural language stories satisfying constraints from natural language critiques or preferences.
We train a contrastive bi-encoder model to align stories with human critiques, building a general purpose preference model.
We further fine-tune the contrastive reward model using a prompt-learning technique to increase story generation robustness.
arXiv Detail & Related papers (2022-10-14T13:21:33Z) - DialogVED: A Pre-trained Latent Variable Encoder-Decoder Model for
Dialog Response Generation [80.45816053153722]
DialogVED introduces continuous latent variables into the enhanced encoder-decoder pre-training framework to increase the relevance and diversity of responses.
We conduct experiments on PersonaChat, DailyDialog, and DSTC7-AVSD benchmarks for response generation.
arXiv Detail & Related papers (2022-04-27T16:18:15Z) - A Model-Agnostic Data Manipulation Method for Persona-based Dialogue
Generation [107.82729587882397]
It is expensive to scale up current persona-based dialogue datasets.
Each data sample in this task is more complex to learn with than conventional dialogue data.
We propose a data manipulation method, which is model-agnostic to be packed with any persona-based dialogue generation model.
arXiv Detail & Related papers (2022-04-21T03:49:54Z) - Multimodal Emotion Recognition using Transfer Learning from Speaker
Recognition and BERT-based models [53.31917090073727]
We propose a neural network-based emotion recognition framework that uses a late fusion of transfer-learned and fine-tuned models from speech and text modalities.
We evaluate the effectiveness of our proposed multimodal approach on the interactive emotional dyadic motion capture dataset.
arXiv Detail & Related papers (2022-02-16T00:23:42Z) - Context Matters in Semantically Controlled Language Generation for
Task-oriented Dialogue Systems [6.1478669848771546]
This work combines information about the dialogue history encoded by pre-trained model with a meaning representation of the current system utterance to realize contextual language generation in task-oriented dialogues.
We utilize the pre-trained multi-context ConveRT model for context representation in a model trained from scratch; and leverage the immediate preceding user utterance for context generation in a model adapted from the pre-trained GPT-2.
arXiv Detail & Related papers (2021-11-28T11:48:02Z) - A Taxonomy of Empathetic Response Intents in Human Social Conversations [1.52292571922932]
Open-domain conversational agents are becoming increasingly popular in the natural language processing community.
One of the challenges is enabling them to converse in an empathetic manner.
Current neural response generation methods rely solely on end-to-end learning from large scale conversation data to generate dialogues.
Recent work has shown the promise of combining dialogue act/intent modelling and neural response generation.
arXiv Detail & Related papers (2020-12-07T21:56:45Z) - The Adapter-Bot: All-In-One Controllable Conversational Model [66.48164003532484]
We propose a dialogue model that uses a fixed backbone model such as DialGPT and triggers on-demand dialogue skills via different adapters.
Depending on the skills, the model is able to process multiple knowledge types, such as text, tables, and emphatic responses.
We evaluate our model using automatic evaluation by comparing it with existing state-of-the-art conversational models.
arXiv Detail & Related papers (2020-08-28T10:59:31Z) - An Empirical Investigation of Pre-Trained Transformer Language Models
for Open-Domain Dialogue Generation [23.343006562849126]
We present an empirical investigation of pre-trained Transformer-based auto-regressive language models for the task of open-domain dialogue generation.
Training paradigm of pre-training and fine-tuning is employed to conduct learning.
Experiments are conducted on the typical single-turn and multi-turn dialogue corpora such as Weibo, Douban, Reddit, DailyDialog, and Persona-Chat.
arXiv Detail & Related papers (2020-03-09T15:20:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.