Improved Goal Oriented Dialogue via Utterance Generation and Look Ahead
- URL: http://arxiv.org/abs/2110.12412v1
- Date: Sun, 24 Oct 2021 11:12:48 GMT
- Title: Improved Goal Oriented Dialogue via Utterance Generation and Look Ahead
- Authors: Eyal Ben-David and Boaz Carmeli and Ateret Anaby-Tavor
- Abstract summary: intent prediction can be improved by training a deep text-to-text neural model to generate successive user utterances from unlabeled dialogue data.
We present a novel look-ahead approach that uses user utterance generation to improve intent prediction in time.
- Score: 5.062869359266078
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Goal oriented dialogue systems have become a prominent customer-care
interaction channel for most businesses. However, not all interactions are
smooth, and customer intent misunderstanding is a major cause of dialogue
failure. We show that intent prediction can be improved by training a deep
text-to-text neural model to generate successive user utterances from unlabeled
dialogue data. For that, we define a multi-task training regime that utilizes
successive user-utterance generation to improve the intent prediction. Our
approach achieves the reported improvement due to two complementary factors:
First, it uses a large amount of unlabeled dialogue data for an auxiliary
generation task. Second, it uses the generated user utterance as an additional
signal for the intent prediction model. Lastly, we present a novel look-ahead
approach that uses user utterance generation to improve intent prediction in
inference time. Specifically, we generate counterfactual successive user
utterances for conversations with ambiguous predicted intents, and disambiguate
the prediction by reassessing the concatenated sequence of available and
generated utterances.
Related papers
- A Stack-Propagation Framework for Low-Resource Personalized Dialogue Generation [29.348053519918928]
We propose a novel stack-propagation framework for learning a dialogue generation and understanding pipeline.
The proposed framework can benefit from the stacked encoder and decoders to learn from much smaller personalized dialogue data.
arXiv Detail & Related papers (2024-10-26T13:09:21Z) - Leveraging Implicit Feedback from Deployment Data in Dialogue [83.02878726357523]
We study improving social conversational agents by learning from natural dialogue between users and a deployed model.
We leverage signals like user response length, sentiment and reaction of the future human utterances in the collected dialogue episodes.
arXiv Detail & Related papers (2023-07-26T11:34:53Z) - Pre-training Multi-party Dialogue Models with Latent Discourse Inference [85.9683181507206]
We pre-train a model that understands the discourse structure of multi-party dialogues, namely, to whom each utterance is replying.
To fully utilize the unlabeled data, we propose to treat the discourse structures as latent variables, then jointly infer them and pre-train the discourse-aware model.
arXiv Detail & Related papers (2023-05-24T14:06:27Z) - EM Pre-training for Multi-party Dialogue Response Generation [86.25289241604199]
In multi-party dialogues, the addressee of a response utterance should be specified before it is generated.
We propose an Expectation-Maximization (EM) approach that iteratively performs the expectation steps to generate addressee labels.
arXiv Detail & Related papers (2023-05-21T09:22:41Z) - Controllable Mixed-Initiative Dialogue Generation through Prompting [50.03458333265885]
Mixed-initiative dialogue tasks involve repeated exchanges of information and conversational control.
Agents gain control by generating responses that follow particular dialogue intents or strategies, prescribed by a policy planner.
Standard approach has been fine-tuning pre-trained language models to perform generation conditioned on these intents.
We instead prompt large language models as a drop-in replacement to fine-tuning on conditional generation.
arXiv Detail & Related papers (2023-05-06T23:11:25Z) - Improving a sequence-to-sequence nlp model using a reinforcement
learning policy algorithm [0.0]
Current neural network models of dialogue generation show great promise for generating answers for chatty agents.
But they are short-sighted in that they predict utterances one at a time while disregarding their impact on future outcomes.
This work commemorates a preliminary step toward developing a neural conversational model based on the long-term success of dialogues.
arXiv Detail & Related papers (2022-12-28T22:46:57Z) - Effective user intent mining with unsupervised word representation
models and topic modelling [0.0]
The explosion of e-commerce has led to a significant increase in text conversation between customers and agents.
We propose an approach to data mining the conversation intents behind the textual data.
arXiv Detail & Related papers (2021-09-04T01:52:12Z) - CloneBot: Personalized Dialogue-Response Predictions [0.0]
The project task was to create a model that, given a speaker ID, chat history, and an utterance query, can predict the response utterance in a conversation.
The model is personalized for each speaker. This task can be a useful tool for building speech bots that talk in a human-like manner in a live conversation.
arXiv Detail & Related papers (2021-03-31T01:15:37Z) - Predict-then-Decide: A Predictive Approach for Wait or Answer Task in
Dialogue Systems [24.560203199376478]
We propose a predictive approach named Predict-then-Decide (PTD) to tackle this Wait-or-Answer problem.
We conduct experiments on two real-life scenarios and three public datasets.
arXiv Detail & Related papers (2020-05-27T01:48:54Z) - Learning an Unreferenced Metric for Online Dialogue Evaluation [53.38078951628143]
We propose an unreferenced automated evaluation metric that uses large pre-trained language models to extract latent representations of utterances.
We show that our model achieves higher correlation with human annotations in an online setting, while not requiring true responses for comparison during inference.
arXiv Detail & Related papers (2020-05-01T20:01:39Z) - TOD-BERT: Pre-trained Natural Language Understanding for Task-Oriented
Dialogue [113.45485470103762]
In this work, we unify nine human-human and multi-turn task-oriented dialogue datasets for language modeling.
To better model dialogue behavior during pre-training, we incorporate user and system tokens into the masked language modeling.
arXiv Detail & Related papers (2020-04-15T04:09:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.