Pneg: Prompt-based Negative Response Generation for Dialogue Response
Selection Task
- URL: http://arxiv.org/abs/2210.17238v1
- Date: Mon, 31 Oct 2022 11:49:49 GMT
- Title: Pneg: Prompt-based Negative Response Generation for Dialogue Response
Selection Task
- Authors: Nyoungwoo Lee, ChaeHun Park, Ho-Jin Choi, and Jaegul Choo
- Abstract summary: In retrieval-based dialogue systems, a response selection model acts as a ranker to select the most appropriate response among several candidates.
Recent studies have shown that leveraging adversarial responses as negative training samples is useful for improving the discriminating power of the selection model.
This paper proposes a simple but efficient method for generating adversarial negative responses leveraging a large-scale language model.
- Score: 27.513992470527427
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In retrieval-based dialogue systems, a response selection model acts as a
ranker to select the most appropriate response among several candidates.
However, such selection models tend to rely on context-response content
similarity, which makes models vulnerable to adversarial responses that are
semantically similar but not relevant to the dialogue context. Recent studies
have shown that leveraging these adversarial responses as negative training
samples is useful for improving the discriminating power of the selection
model. Nevertheless, collecting human-written adversarial responses is
expensive, and existing synthesizing methods often have limited scalability. To
overcome these limitations, this paper proposes a simple but efficient method
for generating adversarial negative responses leveraging a large-scale language
model. Experimental results on dialogue selection tasks show that our method
outperforms other methods of synthesizing adversarial negative responses. These
results suggest that our method can be an effective alternative to human
annotators in generating adversarial responses. Our dataset and generation code
is available at https://github.com/leenw23/generating-negatives-by-gpt3.
Related papers
- JoTR: A Joint Transformer and Reinforcement Learning Framework for
Dialog Policy Learning [53.83063435640911]
Dialogue policy learning (DPL) is a crucial component of dialogue modelling.
We introduce a novel framework, JoTR, to generate flexible dialogue actions.
Unlike traditional methods, JoTR formulates a word-level policy that allows for a more dynamic and adaptable dialogue action generation.
arXiv Detail & Related papers (2023-09-01T03:19:53Z) - SimOAP: Improve Coherence and Consistency in Persona-based Dialogue
Generation via Over-sampling and Post-evaluation [54.66399120084227]
Language models trained on large-scale corpora can generate remarkably fluent results in open-domain dialogue.
For the persona-based dialogue generation task, consistency and coherence are great challenges for language models.
A two-stage SimOAP strategy is proposed, i.e., over-sampling and post-evaluation.
arXiv Detail & Related papers (2023-05-18T17:23:00Z) - Reranking Overgenerated Responses for End-to-End Task-Oriented Dialogue
Systems [71.33737787564966]
End-to-end (E2E) task-oriented dialogue (ToD) systems are prone to fall into the so-called 'likelihood trap'
We propose a reranking method which aims to select high-quality items from the lists of responses initially overgenerated by the system.
Our methods improve a state-of-the-art E2E ToD system by 2.4 BLEU, 3.2 ROUGE, and 2.8 METEOR scores, achieving new peak results.
arXiv Detail & Related papers (2022-11-07T15:59:49Z) - A Systematic Evaluation of Response Selection for Open Domain Dialogue [36.88551817451512]
We curated a dataset where responses from multiple response generators produced for the same dialog context are manually annotated as appropriate (positive) and inappropriate (negative)
We conduct a systematic evaluation of state-of-the-art methods for response selection, and demonstrate that both strategies of using multiple positive candidates and using manually verified hard negative candidates can bring in significant performance improvement in comparison to using the adversarial training data, e.g., increase of 3% and 13% in Recall@1 score, respectively.
arXiv Detail & Related papers (2022-08-08T19:33:30Z) - Diversifying Neural Dialogue Generation via Negative Distillation [11.124375734351826]
We propose a novel negative training paradigm, called negative distillation, to keep the model away from the undesirable generic responses.
Empirical results show that our method outperforms previous negative training methods significantly.
arXiv Detail & Related papers (2022-05-05T17:14:56Z) - Challenging Instances are Worth Learning: Generating Valuable Negative
Samples for Response Selection Training [16.34984384383166]
A response selection module is usually trained on the annotated positive response and sampled negative responses.
We employ pre-trained language models to construct more challenging negative instances to enhance the model robustness.
Our method brings significant and stable improvements on the dialogue response selection capacity.
arXiv Detail & Related papers (2021-09-14T09:16:24Z) - Synthesizing Adversarial Negative Responses for Robust Response Ranking
and Evaluation [34.52276336319678]
Open-domain neural dialogue models have achieved high performance in response ranking and evaluation tasks.
Over-reliance on content similarity makes the models less sensitive to the presence of inconsistencies.
We propose approaches for automatically creating adversarial negative training data.
arXiv Detail & Related papers (2021-06-10T16:20:55Z) - Generating Dialogue Responses from a Semantic Latent Space [75.18449428414736]
We propose an alternative to the end-to-end classification on vocabulary.
We learn the pair relationship between the prompts and responses as a regression task on a latent space.
Human evaluation showed that learning the task on a continuous space can generate responses that are both relevant and informative.
arXiv Detail & Related papers (2020-10-04T19:06:16Z) - Learning an Effective Context-Response Matching Model with
Self-Supervised Tasks for Retrieval-based Dialogues [88.73739515457116]
We introduce four self-supervised tasks including next session prediction, utterance restoration, incoherence detection and consistency discrimination.
We jointly train the PLM-based response selection model with these auxiliary tasks in a multi-task manner.
Experiment results indicate that the proposed auxiliary self-supervised tasks bring significant improvement for multi-turn response selection.
arXiv Detail & Related papers (2020-09-14T08:44:46Z) - Weakly-Supervised Neural Response Selection from an Ensemble of
Task-Specialised Dialogue Agents [11.21333474984984]
We model the problem of selecting the best response from a set of responses generated by a heterogeneous set of dialogue agents.
The proposed method is trained to predict a coherent set of responses within a single conversation, considering its own predictions via a curriculum training mechanism.
arXiv Detail & Related papers (2020-05-06T18:40:26Z) - Counterfactual Off-Policy Training for Neural Response Generation [94.76649147381232]
We propose to explore potential responses by counterfactual reasoning.
Training on the counterfactual responses under the adversarial learning framework helps to explore the high-reward area of the potential response space.
An empirical study on the DailyDialog dataset shows that our approach significantly outperforms the HRED model.
arXiv Detail & Related papers (2020-04-29T22:46:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.