Improving Response Quality with Backward Reasoning in Open-domain
Dialogue Systems
- URL: http://arxiv.org/abs/2105.00079v1
- Date: Fri, 30 Apr 2021 20:38:27 GMT
- Title: Improving Response Quality with Backward Reasoning in Open-domain
Dialogue Systems
- Authors: Ziming Li, Julia Kiseleva, Maarten de Rijke
- Abstract summary: We propose to train the generation model in a bidirectional manner by adding a backward reasoning step to the vanilla encoder-decoder training.
The proposed backward reasoning step pushes the model to produce more informative and coherent content.
Our method can improve response quality without introducing side information.
- Score: 53.160025961101354
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Being able to generate informative and coherent dialogue responses is crucial
when designing human-like open-domain dialogue systems. Encoder-decoder-based
dialogue models tend to produce generic and dull responses during the decoding
step because the most predictable response is likely to be a non-informative
response instead of the most suitable one. To alleviate this problem, we
propose to train the generation model in a bidirectional manner by adding a
backward reasoning step to the vanilla encoder-decoder training. The proposed
backward reasoning step pushes the model to produce more informative and
coherent content because the forward generation step's output is used to infer
the dialogue context in the backward direction. The advantage of our method is
that the forward generation and backward reasoning steps are trained
simultaneously through the use of a latent variable to facilitate bidirectional
optimization. Our method can improve response quality without introducing side
information (e.g., a pre-trained topic model). The proposed bidirectional
response generation method achieves state-of-the-art performance for response
quality.
Related papers
- A Stack-Propagation Framework for Low-Resource Personalized Dialogue Generation [29.348053519918928]
We propose a novel stack-propagation framework for learning a dialogue generation and understanding pipeline.
The proposed framework can benefit from the stacked encoder and decoders to learn from much smaller personalized dialogue data.
arXiv Detail & Related papers (2024-10-26T13:09:21Z) - PICK: Polished & Informed Candidate Scoring for Knowledge-Grounded
Dialogue Systems [59.1250765143521]
Current knowledge-grounded dialogue systems often fail to align the generated responses with human-preferred qualities.
We propose Polished & Informed Candidate Scoring (PICK), a generation re-scoring framework.
We demonstrate the effectiveness of PICK in generating responses that are more faithful while keeping them relevant to the dialogue history.
arXiv Detail & Related papers (2023-09-19T08:27:09Z) - Promoting Open-domain Dialogue Generation through Learning Pattern
Information between Contexts and Responses [5.936682548344234]
This paper improves the quality of generated responses by learning the implicit pattern information between contexts and responses in the training samples.
We also design a response-aware mechanism for mining the implicit pattern information between contexts and responses so that the generated replies are more diverse and approximate to human replies.
arXiv Detail & Related papers (2023-09-06T08:11:39Z) - Diverse and Faithful Knowledge-Grounded Dialogue Generation via
Sequential Posterior Inference [82.28542500317445]
We present an end-to-end learning framework, termed Sequential Posterior Inference (SPI), capable of selecting knowledge and generating dialogues.
Unlike other methods, SPI does not require the inference network or assume a simple geometry of the posterior distribution.
arXiv Detail & Related papers (2023-06-01T21:23:13Z) - EM Pre-training for Multi-party Dialogue Response Generation [86.25289241604199]
In multi-party dialogues, the addressee of a response utterance should be specified before it is generated.
We propose an Expectation-Maximization (EM) approach that iteratively performs the expectation steps to generate addressee labels.
arXiv Detail & Related papers (2023-05-21T09:22:41Z) - DialogVED: A Pre-trained Latent Variable Encoder-Decoder Model for
Dialog Response Generation [80.45816053153722]
DialogVED introduces continuous latent variables into the enhanced encoder-decoder pre-training framework to increase the relevance and diversity of responses.
We conduct experiments on PersonaChat, DailyDialog, and DSTC7-AVSD benchmarks for response generation.
arXiv Detail & Related papers (2022-04-27T16:18:15Z) - Achieving Conversational Goals with Unsupervised Post-hoc Knowledge
Injection [37.15893335147598]
A limitation of current neural dialog models is that they tend to suffer from a lack of specificity and informativeness in generated responses.
We propose a post-hoc knowledge-injection technique where we first retrieve a diverse set of relevant knowledge snippets conditioned on both the dialog history and an initial response from an existing dialog model.
We construct multiple candidate responses, individually injecting each retrieved snippet into the initial response using a gradient-based decoding method, and then select the final response with an unsupervised ranking step.
arXiv Detail & Related papers (2022-03-22T00:42:27Z) - A Speaker-aware Parallel Hierarchical Attentive Encoder-Decoder Model
for Multi-turn Dialogue Generation [13.820298189734686]
This paper presents a novel open-domain dialogue generation model emphasizing the differentiation of speakers in multi-turn conversations.
Our empirical results show that PHAED outperforms the state-of-the-art in both automatic and human evaluations.
arXiv Detail & Related papers (2021-10-13T16:08:29Z) - Reasoning in Dialog: Improving Response Generation by Context Reading
Comprehension [49.92173751203827]
In multi-turn dialog, utterances do not always take the full form of sentences.
We propose to improve the response generation performance by examining the model's ability to answer a reading comprehension question.
arXiv Detail & Related papers (2020-12-14T10:58:01Z) - Posterior-GAN: Towards Informative and Coherent Response Generation with
Posterior Generative Adversarial Network [38.576579498740244]
We propose a novel encoder-decoder based generative adversarial learning framework, Posterior Generative Adversarial Network (Posterior-GAN)
Experimental results demonstrate that our method effectively boosts the informativeness and coherence of the generated response on both automatic and human evaluation.
arXiv Detail & Related papers (2020-03-04T11:57:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.