Self-Adversarial Learning with Comparative Discrimination for Text
Generation
- URL: http://arxiv.org/abs/2001.11691v2
- Date: Wed, 12 Feb 2020 09:18:24 GMT
- Title: Self-Adversarial Learning with Comparative Discrimination for Text
Generation
- Authors: Wangchunshu Zhou, Tao Ge, Ke Xu, Furu Wei, Ming Zhou
- Abstract summary: We propose a novel self-adversarial learning (SAL) paradigm for improving GANs' performance in text generation.
During training, SAL rewards the generator when its currently generated sentence is found to be better than its previously generated samples.
Experiments on text generation benchmark datasets show that our proposed approach substantially improves both the quality and the diversity.
- Score: 111.18614166615968
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Conventional Generative Adversarial Networks (GANs) for text generation tend
to have issues of reward sparsity and mode collapse that affect the quality and
diversity of generated samples. To address the issues, we propose a novel
self-adversarial learning (SAL) paradigm for improving GANs' performance in
text generation. In contrast to standard GANs that use a binary classifier as
its discriminator to predict whether a sample is real or generated, SAL employs
a comparative discriminator which is a pairwise classifier for comparing the
text quality between a pair of samples. During training, SAL rewards the
generator when its currently generated sentence is found to be better than its
previously generated samples. This self-improvement reward mechanism allows the
model to receive credits more easily and avoid collapsing towards the limited
number of real samples, which not only helps alleviate the reward sparsity
issue but also reduces the risk of mode collapse. Experiments on text
generation benchmark datasets show that our proposed approach substantially
improves both the quality and the diversity, and yields more stable performance
compared to the previous GANs for text generation.
Related papers
- Improving Sampling Methods for Fine-tuning SentenceBERT in Text Streams [49.3179290313959]
This study explores the efficacy of seven text sampling methods designed to selectively fine-tune language models.
We precisely assess the impact of these methods on fine-tuning the SBERT model using four different loss functions.
Our findings indicate that Softmax loss and Batch All Triplets loss are particularly effective for text stream classification.
arXiv Detail & Related papers (2024-03-18T23:41:52Z) - Selectively increasing the diversity of GAN-generated samples [8.980453507536017]
We propose a novel method to selectively increase the diversity of GAN-generated samples.
We show the superiority of our method in a synthetic benchmark as well as a real-life scenario simulating data from the Zero Degree Calorimeter of ALICE experiment in CERN.
arXiv Detail & Related papers (2022-07-04T16:27:06Z) - ReSmooth: Detecting and Utilizing OOD Samples when Training with Data
Augmentation [57.38418881020046]
Recent DA techniques always meet the need for diversity in augmented training samples.
An augmentation strategy that has a high diversity usually introduces out-of-distribution (OOD) augmented samples.
We propose ReSmooth, a framework that firstly detects OOD samples in augmented samples and then leverages them.
arXiv Detail & Related papers (2022-05-25T09:29:27Z) - A Well-Composed Text is Half Done! Composition Sampling for Diverse
Conditional Generation [79.98319703471596]
We propose Composition Sampling, a simple but effective method to generate diverse outputs for conditional generation of higher quality.
It builds on recently proposed plan-based neural generation models that are trained to first create a composition of the output and then generate by conditioning on it and the input.
arXiv Detail & Related papers (2022-03-28T21:24:03Z) - Energy-bounded Learning for Robust Models of Code [16.592638312365164]
In programming, learning code representations has a variety of applications, including code classification, code search, comment generation, bug prediction, and so on.
We propose the use of an energy-bounded learning objective function to assign a higher score to in-distribution samples and a lower score to out-of-distribution samples in order to incorporate such out-of-distribution samples into the training process of source code models.
arXiv Detail & Related papers (2021-12-20T06:28:56Z) - Discriminatively-Tuned Generative Classifiers for Robust Natural
Language Inference [59.62779187457773]
We propose a generative classifier for natural language inference (NLI)
We compare it to five baselines, including discriminative models and large-scale pretrained language representation models like BERT.
Experiments show that GenNLI outperforms both discriminative and pretrained baselines across several challenging NLI experimental settings.
arXiv Detail & Related papers (2020-10-08T04:44:00Z) - TextGAIL: Generative Adversarial Imitation Learning for Text Generation [68.3579946817937]
We propose a generative adversarial imitation learning framework for text generation that uses large pre-trained language models to provide more reliable reward guidance.
Our approach uses contrastive discriminator, and proximal policy optimization (PPO) to stabilize and improve text generation performance.
arXiv Detail & Related papers (2020-04-07T00:24:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.