Challenging Instances are Worth Learning: Generating Valuable Negative
Samples for Response Selection Training
- URL: http://arxiv.org/abs/2109.06538v1
- Date: Tue, 14 Sep 2021 09:16:24 GMT
- Title: Challenging Instances are Worth Learning: Generating Valuable Negative
Samples for Response Selection Training
- Authors: Yao Qiu, Jinchao Zhang, Huiying Ren, Jie Zhou
- Abstract summary: A response selection module is usually trained on the annotated positive response and sampled negative responses.
We employ pre-trained language models to construct more challenging negative instances to enhance the model robustness.
Our method brings significant and stable improvements on the dialogue response selection capacity.
- Score: 16.34984384383166
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Retrieval-based chatbot selects the appropriate response from candidates
according to the context, which heavily depends on a response selection module.
A response selection module is generally a scoring model to evaluate candidates
and is usually trained on the annotated positive response and sampled negative
responses. Sampling negative responses lead to two risks: a). The sampled
negative instances, especially that from random sampling methods, are mostly
irrelevant to the dialogue context and too easy to be fitted at the training
stage while causing a weak model in the real scenario. b). The so-called
negative instances may be positive, which is known as the fake negative
problem. To address the above issue, we employ pre-trained language models,
such as the DialoGPT to construct more challenging negative instances to
enhance the model robustness. Specifically, we provide garbled context to the
pre-trained model to generate responses and filter the fake negative ones. In
this way, our negative instances are fluent, context-related, and more
challenging for the model to learn, while can not be positive. Extensive
experiments show that our method brings significant and stable improvements on
the dialogue response selection capacity.
Related papers
- Generating Enhanced Negatives for Training Language-Based Object Detectors [86.1914216335631]
We propose to leverage the vast knowledge built into modern generative models to automatically build negatives that are more relevant to the original data.
Specifically, we use large-language-models to generate negative text descriptions, and text-to-image diffusion models to also generate corresponding negative images.
Our experimental analysis confirms the relevance of the generated negative data, and its use in language-based detectors improves performance on two complex benchmarks.
arXiv Detail & Related papers (2023-12-29T23:04:00Z) - Your Negative May not Be True Negative: Boosting Image-Text Matching
with False Negative Elimination [62.18768931714238]
We propose a novel False Negative Elimination (FNE) strategy to select negatives via sampling.
The results demonstrate the superiority of our proposed false negative elimination strategy.
arXiv Detail & Related papers (2023-08-08T16:31:43Z) - SCENE: Self-Labeled Counterfactuals for Extrapolating to Negative
Examples [23.77077091225583]
Self-labeled Counterfactuals for Extrapolating to Negative Examples (SCENE) is an automatic method for synthesizing training data.
With access to only answerable training examples, SCENE can close 69.6% of the performance gap on SQuAD 2.0.
arXiv Detail & Related papers (2023-05-13T19:30:58Z) - Language Model Pre-training on True Negatives [109.73819321246062]
Discriminative pre-trained language models (PLMs) learn to predict original texts from intentionally corrupted ones.
Existing PLMs simply treat all corrupted texts as equal negative without any examination.
We design enhanced pre-training methods to counteract false negative predictions and encourage pre-training language models on true negatives.
arXiv Detail & Related papers (2022-12-01T12:24:19Z) - Pneg: Prompt-based Negative Response Generation for Dialogue Response
Selection Task [27.513992470527427]
In retrieval-based dialogue systems, a response selection model acts as a ranker to select the most appropriate response among several candidates.
Recent studies have shown that leveraging adversarial responses as negative training samples is useful for improving the discriminating power of the selection model.
This paper proposes a simple but efficient method for generating adversarial negative responses leveraging a large-scale language model.
arXiv Detail & Related papers (2022-10-31T11:49:49Z) - SimANS: Simple Ambiguous Negatives Sampling for Dense Text Retrieval [126.22182758461244]
We show that according to the measured relevance scores, the negatives ranked around the positives are generally more informative and less likely to be false negatives.
We propose a simple ambiguous negatives sampling method, SimANS, which incorporates a new sampling probability distribution to sample more ambiguous negatives.
arXiv Detail & Related papers (2022-10-21T07:18:05Z) - A Systematic Evaluation of Response Selection for Open Domain Dialogue [36.88551817451512]
We curated a dataset where responses from multiple response generators produced for the same dialog context are manually annotated as appropriate (positive) and inappropriate (negative)
We conduct a systematic evaluation of state-of-the-art methods for response selection, and demonstrate that both strategies of using multiple positive candidates and using manually verified hard negative candidates can bring in significant performance improvement in comparison to using the adversarial training data, e.g., increase of 3% and 13% in Recall@1 score, respectively.
arXiv Detail & Related papers (2022-08-08T19:33:30Z) - Diversifying Neural Dialogue Generation via Negative Distillation [11.124375734351826]
We propose a novel negative training paradigm, called negative distillation, to keep the model away from the undesirable generic responses.
Empirical results show that our method outperforms previous negative training methods significantly.
arXiv Detail & Related papers (2022-05-05T17:14:56Z) - Rethinking InfoNCE: How Many Negative Samples Do You Need? [54.146208195806636]
We study how many negative samples are optimal for InfoNCE in different scenarios via a semi-quantitative theoretical framework.
We estimate the optimal negative sampling ratio using the $K$ value that maximizes the training effectiveness function.
arXiv Detail & Related papers (2021-05-27T08:38:29Z) - Contrastive Learning with Hard Negative Samples [80.12117639845678]
We develop a new family of unsupervised sampling methods for selecting hard negative samples.
A limiting case of this sampling results in a representation that tightly clusters each class, and pushes different classes as far apart as possible.
The proposed method improves downstream performance across multiple modalities, requires only few additional lines of code to implement, and introduces no computational overhead.
arXiv Detail & Related papers (2020-10-09T14:18:53Z) - Group-wise Contrastive Learning for Neural Dialogue Generation [29.749195182401344]
We introduce contrastive learning into dialogue generation, where the model explicitly perceives the difference between the well-chosen positive and negative utterances.
To manage the multi-mapping relations prevailed in human conversation, we augment contrastive dialogue learning with group-wise dual sampling.
arXiv Detail & Related papers (2020-09-16T08:28:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.