Towards Efficiently Diversifying Dialogue Generation via Embedding
Augmentation
- URL: http://arxiv.org/abs/2103.01534v1
- Date: Tue, 2 Mar 2021 07:28:56 GMT
- Title: Towards Efficiently Diversifying Dialogue Generation via Embedding
Augmentation
- Authors: Yu Cao, Liang Ding, Zhiliang Tian, Meng Fang
- Abstract summary: We propose to promote the generation diversity of the neural dialogue models via soft embedding augmentation.
New embeddings serve as the input of the model to replace the original one.
Our experimental results on two datasets illustrate that our proposed method is capable of generating more diverse responses than raw models.
- Score: 24.940159321142165
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Dialogue generation models face the challenge of producing generic and
repetitive responses. Unlike previous augmentation methods that mostly focus on
token manipulation and ignore the essential variety within a single sample
using hard labels, we propose to promote the generation diversity of the neural
dialogue models via soft embedding augmentation along with soft labels in this
paper. Particularly, we select some key input tokens and fuse their embeddings
together with embeddings from their semantic-neighbor tokens. The new
embeddings serve as the input of the model to replace the original one.
Besides, soft labels are used in loss calculation, resulting in multi-target
supervision for a given input. Our experimental results on two datasets
illustrate that our proposed method is capable of generating more diverse
responses than raw models while remains a similar n-gram accuracy that ensures
the quality of generated responses.
Related papers
- Vector-Quantized Prompt Learning for Paraphrase Generation [18.40940464497253]
This paper proposes to generate diverse and high-quality paraphrases by exploiting the pre-trained models with instance-dependent prompts.
Extensive experiments demonstrate that the proposed method achieves new state-of-art results on three benchmark datasets.
arXiv Detail & Related papers (2023-11-25T07:13:06Z) - Improving the Robustness of Summarization Systems with Dual Augmentation [68.53139002203118]
A robust summarization system should be able to capture the gist of the document, regardless of the specific word choices or noise in the input.
We first explore the summarization models' robustness against perturbations including word-level synonym substitution and noise.
We propose a SummAttacker, which is an efficient approach to generating adversarial samples based on language models.
arXiv Detail & Related papers (2023-06-01T19:04:17Z) - Self-Evolution Learning for Mixup: Enhance Data Augmentation on Few-Shot
Text Classification Tasks [75.42002070547267]
We propose a self evolution learning (SE) based mixup approach for data augmentation in text classification.
We introduce a novel instance specific label smoothing approach, which linearly interpolates the model's output and one hot labels of the original samples to generate new soft for label mixing up.
arXiv Detail & Related papers (2023-05-22T23:43:23Z) - A Contrastive Framework for Neural Text Generation [46.845997620234265]
We show that an underlying reason for model degeneration is the anisotropic distribution of token representations.
We present a contrastive solution: (i) SimCTG, a contrastive training objective to calibrate the model's representation space, and (ii) a decoding method -- contrastive search -- to encourage diversity while maintaining coherence in the generated text.
arXiv Detail & Related papers (2022-02-13T21:46:14Z) - Diversifying Dialog Generation via Adaptive Label Smoothing [49.216146120454745]
We propose an Adaptive Label Smoothing (AdaLabel) approach that can adaptively estimate a target label distribution at each time step for different contexts.
Our approach outperforms various competitive baselines in producing diverse responses.
arXiv Detail & Related papers (2021-05-30T14:41:09Z) - Improve Variational Autoencoder for Text Generationwith Discrete Latent
Bottleneck [52.08901549360262]
Variational autoencoders (VAEs) are essential tools in end-to-end representation learning.
VAEs tend to ignore latent variables with a strong auto-regressive decoder.
We propose a principled approach to enforce an implicit latent feature matching in a more compact latent space.
arXiv Detail & Related papers (2020-04-22T14:41:37Z) - Generating diverse and natural text-to-speech samples using a quantized
fine-grained VAE and auto-regressive prosody prior [53.69310441063162]
This paper proposes a sequential prior in a discrete latent space which can generate more naturally sounding samples.
We evaluate the approach using listening tests, objective metrics of automatic speech recognition (ASR) performance, and measurements of prosody attributes.
arXiv Detail & Related papers (2020-02-06T12:35:50Z) - AvgOut: A Simple Output-Probability Measure to Eliminate Dull Responses [97.50616524350123]
We build dialogue models that are dynamically aware of what utterances or tokens are dull without any feature-engineering.
The first model, MinAvgOut, directly maximizes the diversity score through the output distributions of each batch.
The second model, Label Fine-Tuning (LFT), prepends to the source sequence a label continuously scaled by the diversity score to control the diversity level.
The third model, RL, adopts Reinforcement Learning and treats the diversity score as a reward signal.
arXiv Detail & Related papers (2020-01-15T18:32:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.