Unsupervised Contextual Paraphrase Generation using Lexical Control and
Reinforcement Learning
- URL: http://arxiv.org/abs/2103.12777v1
- Date: Tue, 23 Mar 2021 18:22:03 GMT
- Title: Unsupervised Contextual Paraphrase Generation using Lexical Control and
Reinforcement Learning
- Authors: Sonal Garg, Sumanth Prabhu, Hemant Misra, and G. Srinivasaraghavan
- Abstract summary: We propose an unsupervised frame-work to generate contextual paraphrases using autoregressive models.
We also propose an automated metric based on Semantic Similarity, Textual Entailment, Expression Diversity and Fluency to evaluate the quality of contextual paraphrases.
- Score: 3.2811284938530636
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Customer support via chat requires agents to resolve customer queries with
minimum wait time and maximum customer satisfaction. Given that the agents as
well as the customers can have varying levels of literacy, the overall quality
of responses provided by the agents tend to be poor if they are not predefined.
But using only static responses can lead to customer detraction as the
customers tend to feel that they are no longer interacting with a human. Hence,
it is vital to have variations of the static responses to reduce monotonicity
of the responses. However, maintaining a list of such variations can be
expensive. Given the conversation context and the agent response, we propose an
unsupervised frame-work to generate contextual paraphrases using autoregressive
models. We also propose an automated metric based on Semantic Similarity,
Textual Entailment, Expression Diversity and Fluency to evaluate the quality of
contextual paraphrases and demonstrate performance improvement with
Reinforcement Learning (RL) fine-tuning using the automated metric as the
reward function.
Related papers
- RSVP: Customer Intent Detection via Agent Response Contrastive and
Generative Pre-Training [16.183398769581075]
RSVP is a self-supervised framework dedicated to task-oriented dialogues.
It incorporates agent responses for pre-training in a two-stage manner.
Our benchmark results for two real-world customer service datasets show that RSVP significantly outperforms the state-of-the-art baselines.
arXiv Detail & Related papers (2023-10-15T08:21:38Z) - Leveraging Implicit Feedback from Deployment Data in Dialogue [83.02878726357523]
We study improving social conversational agents by learning from natural dialogue between users and a deployed model.
We leverage signals like user response length, sentiment and reaction of the future human utterances in the collected dialogue episodes.
arXiv Detail & Related papers (2023-07-26T11:34:53Z) - Controllable Mixed-Initiative Dialogue Generation through Prompting [50.03458333265885]
Mixed-initiative dialogue tasks involve repeated exchanges of information and conversational control.
Agents gain control by generating responses that follow particular dialogue intents or strategies, prescribed by a policy planner.
Standard approach has been fine-tuning pre-trained language models to perform generation conditioned on these intents.
We instead prompt large language models as a drop-in replacement to fine-tuning on conditional generation.
arXiv Detail & Related papers (2023-05-06T23:11:25Z) - WeaSuL: Weakly Supervised Dialogue Policy Learning: Reward Estimation
for Multi-turn Dialogue [17.663449579168297]
We simulate a dialogue between an agent and a user (modelled similar to an agent with supervised learning objective) to interact with each other.
The agent uses dynamic blocking to generate ranked diverse responses and exploration-exploitation to select among the Top-K responses.
Empirical studies with two benchmarks indicate that our model can significantly out-perform the response quality and lead to a successful conversation.
arXiv Detail & Related papers (2021-08-01T08:00:45Z) - Evaluating Empathetic Chatbots in Customer Service Settings [6.523873187705393]
We show that a blended skills chatbots model that responds to customer queries is more likely to resemble actual human agent response if it is trained to recognize emotion and exhibit appropriate empathy.
For our analysis, we leverage a Twitter customer service dataset containing several million customer->agent dialog examples in customer service contexts from 20 well-known brands.
arXiv Detail & Related papers (2021-01-05T03:34:35Z) - Turn-level Dialog Evaluation with Dialog-level Weak Signals for
Bot-Human Hybrid Customer Service Systems [0.0]
We developed a machine learning approach that quantifies multiple aspects of the success or values in Customer Service contacts, at anytime during the interaction.
We show how it improves Amazon customer service quality in several applications.
arXiv Detail & Related papers (2020-10-25T19:36:23Z) - How To Evaluate Your Dialogue System: Probe Tasks as an Alternative for
Token-level Evaluation Metrics [47.20761880464552]
generative dialogue modeling is widely seen as a language modeling task.
The task demands an agent to have a complex natural language understanding of its input text to carry a meaningful interaction with an user.
The automatic metrics used evaluate the quality of the generated text as a proxy to the holistic interaction of the agent.
arXiv Detail & Related papers (2020-08-24T13:28:35Z) - A Controllable Model of Grounded Response Generation [122.7121624884747]
Current end-to-end neural conversation models inherently lack the flexibility to impose semantic control in the response generation process.
We propose a framework that we call controllable grounded response generation (CGRG)
We show that using this framework, a transformer based model with a novel inductive attention mechanism, trained on a conversation-like Reddit dataset, outperforms strong generation baselines.
arXiv Detail & Related papers (2020-05-01T21:22:08Z) - Learning an Unreferenced Metric for Online Dialogue Evaluation [53.38078951628143]
We propose an unreferenced automated evaluation metric that uses large pre-trained language models to extract latent representations of utterances.
We show that our model achieves higher correlation with human annotations in an online setting, while not requiring true responses for comparison during inference.
arXiv Detail & Related papers (2020-05-01T20:01:39Z) - Improving Factual Consistency Between a Response and Persona Facts [64.30785349238619]
Neural models for response generation produce responses that are semantically plausible but not necessarily factually consistent with facts describing the speaker's persona.
We propose to fine-tune these models by reinforcement learning and an efficient reward function that explicitly captures the consistency between a response and persona facts as well as semantic plausibility.
arXiv Detail & Related papers (2020-04-30T18:08:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.