Data Manipulation: Towards Effective Instance Learning for Neural
Dialogue Generation via Learning to Augment and Reweight
- URL: http://arxiv.org/abs/2004.02594v5
- Date: Thu, 11 Jun 2020 14:01:55 GMT
- Title: Data Manipulation: Towards Effective Instance Learning for Neural
Dialogue Generation via Learning to Augment and Reweight
- Authors: Hengyi Cai, Hongshen Chen, Yonghao Song, Cheng Zhang, Xiaofang Zhao,
Dawei Yin
- Abstract summary: Current state-of-the-art neural dialogue models learn from human conversations following the data-driven paradigm.
Due to the open-ended nature of human conversations, the quality of user-generated training data varies greatly.
We propose a data manipulation framework to proactively reshape the data distribution towards reliable samples.
- Score: 39.199204415979196
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Current state-of-the-art neural dialogue models learn from human
conversations following the data-driven paradigm. As such, a reliable training
corpus is the crux of building a robust and well-behaved dialogue model.
However, due to the open-ended nature of human conversations, the quality of
user-generated training data varies greatly, and effective training samples are
typically insufficient while noisy samples frequently appear. This impedes the
learning of those data-driven neural dialogue models. Therefore, effective
dialogue learning requires not only more reliable learning samples, but also
fewer noisy samples. In this paper, we propose a data manipulation framework to
proactively reshape the data distribution towards reliable samples by
augmenting and highlighting effective learning samples as well as reducing the
effect of inefficient samples simultaneously. In particular, the data
manipulation model selectively augments the training samples and assigns an
importance weight to each instance to reform the training data. Note that, the
proposed data manipulation framework is fully data-driven and learnable. It not
only manipulates training samples to optimize the dialogue generation model,
but also learns to increase its manipulation skills through gradient descent
with validation samples. Extensive experiments show that our framework can
improve the dialogue generation performance with respect to various automatic
evaluation metrics and human judgments.
Related papers
- Distilled Datamodel with Reverse Gradient Matching [74.75248610868685]
We introduce an efficient framework for assessing data impact, comprising offline training and online evaluation stages.
Our proposed method achieves comparable model behavior evaluation while significantly speeding up the process compared to the direct retraining method.
arXiv Detail & Related papers (2024-04-22T09:16:14Z) - Robust Task-Oriented Dialogue Generation with Contrastive Pre-training
and Adversarial Filtering [17.7709632238066]
Data artifacts incentivize machine learning models to learn non-transferable generalizations.
We investigate whether popular datasets such as MultiWOZ contain such data artifacts.
We propose a contrastive learning based framework to encourage the model to ignore these cues and focus on learning generalisable patterns.
arXiv Detail & Related papers (2022-05-20T03:13:02Z) - A Model-Agnostic Data Manipulation Method for Persona-based Dialogue
Generation [107.82729587882397]
It is expensive to scale up current persona-based dialogue datasets.
Each data sample in this task is more complex to learn with than conventional dialogue data.
We propose a data manipulation method, which is model-agnostic to be packed with any persona-based dialogue generation model.
arXiv Detail & Related papers (2022-04-21T03:49:54Z) - Pan More Gold from the Sand: Refining Open-domain Dialogue Training with
Noisy Self-Retrieval Generation [41.928481605188146]
We show existing open-domain dialogue generation methods by memorizing context-response paired data with causal or encode-decode language models underutilize the training data.
Our method performed well on zero-shot experiments, which indicates that our method can be more robust to real-world data.
arXiv Detail & Related papers (2022-01-27T08:02:59Z) - Training Data Leakage Analysis in Language Models [6.843491191969066]
We introduce a methodology that investigates identifying the user content in the training data that could be leaked under a strong and realistic threat model.
We propose two metrics to quantify user-level data leakage by measuring a model's ability to produce unique sentence fragments within training data.
arXiv Detail & Related papers (2021-01-14T00:57:32Z) - Dialogue Distillation: Open-Domain Dialogue Augmentation Using Unpaired
Data [61.71319905364992]
We propose a novel data augmentation method for training open-domain dialogue models by utilizing unpaired data.
A data-level distillation process is first proposed to construct augmented dialogues where both post and response are retrieved from the unpaired data.
A ranking module is employed to filter out low-quality dialogues.
A model-level distillation process is employed to distill a teacher model trained on high-quality paired data to augmented dialogue pairs.
arXiv Detail & Related papers (2020-09-20T13:06:38Z) - Automatic Recall Machines: Internal Replay, Continual Learning and the
Brain [104.38824285741248]
Replay in neural networks involves training on sequential data with memorized samples, which counteracts forgetting of previous behavior caused by non-stationarity.
We present a method where these auxiliary samples are generated on the fly, given only the model that is being trained for the assessed objective.
Instead the implicit memory of learned samples within the assessed model itself is exploited.
arXiv Detail & Related papers (2020-06-22T15:07:06Z) - Low-Resource Knowledge-Grounded Dialogue Generation [74.09352261943913]
We consider knowledge-grounded dialogue generation under a natural assumption that only limited training examples are available.
We devise a disentangled response decoder in order to isolate parameters that depend on knowledge-grounded dialogues from the entire generation model.
With only 1/8 training data, our model can achieve the state-of-the-art performance and generalize well on out-of-domain knowledge.
arXiv Detail & Related papers (2020-02-24T16:20:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.