Pan More Gold from the Sand: Refining Open-domain Dialogue Training with
Noisy Self-Retrieval Generation
- URL: http://arxiv.org/abs/2201.11367v1
- Date: Thu, 27 Jan 2022 08:02:59 GMT
- Title: Pan More Gold from the Sand: Refining Open-domain Dialogue Training with
Noisy Self-Retrieval Generation
- Authors: Yihe Wang, Yitong Li, Yasheng Wang, Fei Mi, Pingyi Zhou, Xin Wang, Jin
Liu, Qun Liu, Xin Jiang
- Abstract summary: We show existing open-domain dialogue generation methods by memorizing context-response paired data with causal or encode-decode language models underutilize the training data.
Our method performed well on zero-shot experiments, which indicates that our method can be more robust to real-world data.
- Score: 41.928481605188146
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Real human conversation data are complicated, heterogeneous, and noisy, from
whom building open-domain dialogue systems remains a challenging task. In fact,
such dialogue data can still contain a wealth of information and knowledge,
however, they are not fully explored. In this paper, we show existing
open-domain dialogue generation methods by memorizing context-response paired
data with causal or encode-decode language models underutilize the training
data. Different from current approaches, using external knowledge, we explore a
retrieval-generation training framework that can increase the usage of training
data by directly considering the heterogeneous and noisy training data as the
"evidence". Experiments over publicly available datasets demonstrate that our
method can help models generate better responses, even such training data are
usually impressed as low-quality data. Such performance gain is comparable with
those improved by enlarging the training set, even better. We also found that
the model performance has a positive correlation with the relevance of the
retrieved evidence. Moreover, our method performed well on zero-shot
experiments, which indicates that our method can be more robust to real-world
data.
Related papers
- Noisy Self-Training with Synthetic Queries for Dense Retrieval [49.49928764695172]
We introduce a novel noisy self-training framework combined with synthetic queries.
Experimental results show that our method improves consistently over existing methods.
Our method is data efficient and outperforms competitive baselines.
arXiv Detail & Related papers (2023-11-27T06:19:50Z) - Hexa: Self-Improving for Knowledge-Grounded Dialogue System [13.293318039036562]
We develop a self-improving method to improve the generative performances of intermediate steps without the ground truth data.
In particular, we propose a novel bootstrapping scheme with a guided prompt and a modified loss function to enhance the diversity of appropriate self-generated responses.
arXiv Detail & Related papers (2023-10-10T08:15:24Z) - Weakly Supervised Data Augmentation Through Prompting for Dialogue
Understanding [103.94325597273316]
We present a novel approach that iterates on augmentation quality by applying weakly-supervised filters.
We evaluate our methods on the emotion and act classification tasks in DailyDialog and the intent classification task in Facebook Multilingual Task-Oriented Dialogue.
For DailyDialog specifically, using 10% of the ground truth data we outperform the current state-of-the-art model which uses 100% of the data.
arXiv Detail & Related papers (2022-10-25T17:01:30Z) - Robust Task-Oriented Dialogue Generation with Contrastive Pre-training
and Adversarial Filtering [17.7709632238066]
Data artifacts incentivize machine learning models to learn non-transferable generalizations.
We investigate whether popular datasets such as MultiWOZ contain such data artifacts.
We propose a contrastive learning based framework to encourage the model to ignore these cues and focus on learning generalisable patterns.
arXiv Detail & Related papers (2022-05-20T03:13:02Z) - A Model-Agnostic Data Manipulation Method for Persona-based Dialogue
Generation [107.82729587882397]
It is expensive to scale up current persona-based dialogue datasets.
Each data sample in this task is more complex to learn with than conventional dialogue data.
We propose a data manipulation method, which is model-agnostic to be packed with any persona-based dialogue generation model.
arXiv Detail & Related papers (2022-04-21T03:49:54Z) - Dialogue Distillation: Open-Domain Dialogue Augmentation Using Unpaired
Data [61.71319905364992]
We propose a novel data augmentation method for training open-domain dialogue models by utilizing unpaired data.
A data-level distillation process is first proposed to construct augmented dialogues where both post and response are retrieved from the unpaired data.
A ranking module is employed to filter out low-quality dialogues.
A model-level distillation process is employed to distill a teacher model trained on high-quality paired data to augmented dialogue pairs.
arXiv Detail & Related papers (2020-09-20T13:06:38Z) - Data Manipulation: Towards Effective Instance Learning for Neural
Dialogue Generation via Learning to Augment and Reweight [39.199204415979196]
Current state-of-the-art neural dialogue models learn from human conversations following the data-driven paradigm.
Due to the open-ended nature of human conversations, the quality of user-generated training data varies greatly.
We propose a data manipulation framework to proactively reshape the data distribution towards reliable samples.
arXiv Detail & Related papers (2020-04-06T12:14:09Z) - Low-Resource Knowledge-Grounded Dialogue Generation [74.09352261943913]
We consider knowledge-grounded dialogue generation under a natural assumption that only limited training examples are available.
We devise a disentangled response decoder in order to isolate parameters that depend on knowledge-grounded dialogues from the entire generation model.
With only 1/8 training data, our model can achieve the state-of-the-art performance and generalize well on out-of-domain knowledge.
arXiv Detail & Related papers (2020-02-24T16:20:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.