Adaptive Bridge between Training and Inference for Dialogue
- URL: http://arxiv.org/abs/2110.11560v1
- Date: Fri, 22 Oct 2021 02:43:27 GMT
- Title: Adaptive Bridge between Training and Inference for Dialogue
- Authors: Haoran Xu, Hainan Zhang, Yanyan Zou, Hongshen Chen, Zhuoye Ding,
Yanyan Lan
- Abstract summary: We propose a novel adaptive switching mechanism, which learns to automatically transit between ground-truth learning and generated learning.
Our method achieves a significant improvement in terms of metric-based evaluation and human evaluation.
- Score: 36.64781557775641
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Although exposure bias has been widely studied in some NLP tasks, it faces
its unique challenges in dialogue response generation, the representative
one-to-various generation scenario. In real human dialogue, there are many
appropriate responses for the same context, not only with different
expressions, but also with different topics. Therefore, due to the much bigger
gap between various ground-truth responses and the generated synthetic
response, exposure bias is more challenging in dialogue generation task. What's
more, as MLE encourages the model to only learn the common words among
different ground-truth responses, but ignores the interesting and specific
parts, exposure bias may further lead to the common response generation
problem, such as "I don't know" and "HaHa?" In this paper, we propose a novel
adaptive switching mechanism, which learns to automatically transit between
ground-truth learning and generated learning regarding the word-level matching
score, such as the cosine similarity. Experimental results on both Chinese STC
dataset and English Reddit dataset, show that our adaptive method achieves a
significant improvement in terms of metric-based evaluation and human
evaluation, as compared with the state-of-the-art exposure bias approaches.
Further analysis on NMT task also shows that our model can achieve a
significant improvement.
Related papers
- Multi-level Adaptive Contrastive Learning for Knowledge Internalization
in Dialogue Generation [37.55417272177113]
Knowledge-grounded dialogue generation aims to incorporate external knowledge to supplement the context.
However, the model often fails to internalize this information into responses in a human-like manner.
We propose a Multi-level Adaptive Contrastive Learning framework that dynamically samples negative examples and subsequently penalizes degeneration behaviors.
arXiv Detail & Related papers (2023-10-13T08:16:27Z) - PICK: Polished & Informed Candidate Scoring for Knowledge-Grounded
Dialogue Systems [59.1250765143521]
Current knowledge-grounded dialogue systems often fail to align the generated responses with human-preferred qualities.
We propose Polished & Informed Candidate Scoring (PICK), a generation re-scoring framework.
We demonstrate the effectiveness of PICK in generating responses that are more faithful while keeping them relevant to the dialogue history.
arXiv Detail & Related papers (2023-09-19T08:27:09Z) - Re-mine, Learn and Reason: Exploring the Cross-modal Semantic
Correlations for Language-guided HOI detection [57.13665112065285]
Human-Object Interaction (HOI) detection is a challenging computer vision task.
We present a framework that enhances HOI detection by incorporating structured text knowledge.
arXiv Detail & Related papers (2023-07-25T14:20:52Z) - DialAug: Mixing up Dialogue Contexts in Contrastive Learning for Robust
Conversational Modeling [3.3578533367912025]
We propose a framework that incorporates augmented versions of a dialogue context into the learning objective.
We show that our proposed augmentation method outperforms previous data augmentation approaches.
arXiv Detail & Related papers (2022-04-15T23:39:41Z) - Towards Robust Online Dialogue Response Generation [62.99904593650087]
We argue that this can be caused by a discrepancy between training and real-world testing.
We propose a hierarchical sampling-based method consisting of both utterance-level sampling and semi-utterance-level sampling.
arXiv Detail & Related papers (2022-03-07T06:51:41Z) - Learning from Perturbations: Diverse and Informative Dialogue Generation
with Inverse Adversarial Training [10.17868476063421]
We propose Inverse Adversarial Training (IAT) algorithm for training neural dialogue systems.
IAT encourages the model to be sensitive to the perturbation in the dialogue history and therefore learning from perturbations.
We show that our approach can better model dialogue history and generate more diverse and consistent responses.
arXiv Detail & Related papers (2021-05-31T17:28:37Z) - Refine and Imitate: Reducing Repetition and Inconsistency in Persuasion
Dialogues via Reinforcement Learning and Human Demonstration [45.14559188965439]
We propose to apply reinforcement learning to refine an MLE-based language model without user simulators.
We distill sentence-level information about repetition, inconsistency and task relevance through rewards.
Experiments show that our model outperforms previous state-of-the-art dialogue models on both automatic metrics and human evaluation results.
arXiv Detail & Related papers (2020-12-31T00:02:51Z) - I like fish, especially dolphins: Addressing Contradictions in Dialogue
Modeling [104.09033240889106]
We introduce the DialoguE COntradiction DEtection task (DECODE) and a new conversational dataset containing both human-human and human-bot contradictory dialogues.
We then compare a structured utterance-based approach of using pre-trained Transformer models for contradiction detection with the typical unstructured approach.
arXiv Detail & Related papers (2020-12-24T18:47:49Z) - Group-wise Contrastive Learning for Neural Dialogue Generation [29.749195182401344]
We introduce contrastive learning into dialogue generation, where the model explicitly perceives the difference between the well-chosen positive and negative utterances.
To manage the multi-mapping relations prevailed in human conversation, we augment contrastive dialogue learning with group-wise dual sampling.
arXiv Detail & Related papers (2020-09-16T08:28:30Z) - Learning an Effective Context-Response Matching Model with
Self-Supervised Tasks for Retrieval-based Dialogues [88.73739515457116]
We introduce four self-supervised tasks including next session prediction, utterance restoration, incoherence detection and consistency discrimination.
We jointly train the PLM-based response selection model with these auxiliary tasks in a multi-task manner.
Experiment results indicate that the proposed auxiliary self-supervised tasks bring significant improvement for multi-turn response selection.
arXiv Detail & Related papers (2020-09-14T08:44:46Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.