To Beam Or Not To Beam: That is a Question of Cooperation for Language
GANs
- URL: http://arxiv.org/abs/2106.06363v1
- Date: Fri, 11 Jun 2021 13:04:42 GMT
- Title: To Beam Or Not To Beam: That is a Question of Cooperation for Language
GANs
- Authors: Thomas Scialom, Paul-Alexis Dray, Sylvain Lamprier, Benjamin
Piwowarski, Jacopo Staiano
- Abstract summary: Language GANs require to be optimized from rewards provided by discriminator networks, via reinforcement learning methods.
We show that our SelfGAN framework, built on this cooperative principle, outperforms Teacher Forcing and obtains state-of-the-art results on two challenging tasks.
- Score: 31.040350519448342
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Due to the discrete nature of words, language GANs require to be optimized
from rewards provided by discriminator networks, via reinforcement learning
methods. This is a much harder setting than for continuous tasks, which enjoy
gradient flows from discriminators to generators, usually leading to dramatic
learning instabilities. However, we claim that this can be solved by making
discriminator and generator networks cooperate to produce output sequences
during training. These cooperative outputs, inherently built to obtain higher
discrimination scores, not only provide denser rewards for training, but also
form a more compact artificial set for discriminator training, hence improving
its accuracy and stability. In this paper, we show that our SelfGAN framework,
built on this cooperative principle, outperforms Teacher Forcing and obtains
state-of-the-art results on two challenging tasks, Summarization and Question
Generation.
Related papers
- Deep Boosting Learning: A Brand-new Cooperative Approach for Image-Text Matching [53.05954114863596]
We propose a brand-new Deep Boosting Learning (DBL) algorithm for image-text matching.
An anchor branch is first trained to provide insights into the data properties.
A target branch is concurrently tasked with more adaptive margin constraints to further enlarge the relative distance between matched and unmatched samples.
arXiv Detail & Related papers (2024-04-28T08:44:28Z) - Structural Adversarial Objectives for Self-Supervised Representation
Learning [19.471586646254373]
We propose objectives that task the discriminator for self-supervised representation learning via additional structural modeling responsibilities.
In combination with an efficient smoothness regularizer imposed on the network, these objectives guide the discriminator to learn to extract informative representations.
Experiments demonstrate that equipping GANs with our self-supervised objectives suffices to produce discriminators which, evaluated in terms of representation learning, compete with networks trained by contrastive learning approaches.
arXiv Detail & Related papers (2023-09-30T12:27:53Z) - Improving GANs with A Dynamic Discriminator [106.54552336711997]
We argue that a discriminator with an on-the-fly adjustment on its capacity can better accommodate such a time-varying task.
A comprehensive empirical study confirms that the proposed training strategy, termed as DynamicD, improves the synthesis performance without incurring any additional cost or training objectives.
arXiv Detail & Related papers (2022-09-20T17:57:33Z) - Latent Policies for Adversarial Imitation Learning [21.105328282702885]
This paper considers learning robot locomotion and manipulation tasks from expert demonstrations.
Generative adversarial imitation learning (GAIL) trains a discriminator that distinguishes expert from agent transitions, and in turn use a reward defined by the discriminator output to optimize a policy generator for the agent.
A key insight of this work is that performing imitation learning in a suitable latent task space makes the training process stable, even in challenging high-dimensional problems.
arXiv Detail & Related papers (2022-06-22T18:06:26Z) - Generative Cooperative Networks for Natural Language Generation [25.090455367573988]
We introduce Generative Cooperative Networks, in which the discriminator architecture is cooperatively used along with the generation policy to output samples of realistic texts.
We give theoretical guarantees of convergence for our approach, and study various efficient decoding schemes to empirically achieve state-of-the-art results in two main NLG tasks.
arXiv Detail & Related papers (2022-01-28T18:36:57Z) - Hybrid Generative-Contrastive Representation Learning [32.84066504783469]
We show that a transformer-based encoder-decoder architecture trained with both contrastive and generative losses can learn highly discriminative and robust representations without hurting the generative performance.
arXiv Detail & Related papers (2021-06-11T04:23:48Z) - Training GANs with Stronger Augmentations via Contrastive Discriminator [80.8216679195]
We introduce a contrastive representation learning scheme into the GAN discriminator, coined ContraD.
This "fusion" enables the discriminators to work with much stronger augmentations without increasing their training instability.
Our experimental results show that GANs with ContraD consistently improve FID and IS compared to other recent techniques incorporating data augmentations.
arXiv Detail & Related papers (2021-03-17T16:04:54Z) - Discriminatively-Tuned Generative Classifiers for Robust Natural
Language Inference [59.62779187457773]
We propose a generative classifier for natural language inference (NLI)
We compare it to five baselines, including discriminative models and large-scale pretrained language representation models like BERT.
Experiments show that GenNLI outperforms both discriminative and pretrained baselines across several challenging NLI experimental settings.
arXiv Detail & Related papers (2020-10-08T04:44:00Z) - Bridging the Imitation Gap by Adaptive Insubordination [88.35564081175642]
We show that when the teaching agent makes decisions with access to privileged information, this information is marginalized during imitation learning.
We propose 'Adaptive Insubordination' (ADVISOR) to address this gap.
ADVISOR dynamically weights imitation and reward-based reinforcement learning losses during training, enabling on-the-fly switching between imitation and exploration.
arXiv Detail & Related papers (2020-07-23T17:59:57Z) - Making Method of Moments Great Again? -- How can GANs learn
distributions [34.91089650516183]
Generative Adrial Networks (GANs) are widely used models to learn complex real-world distributions.
In GANs, the training of the generator usually stops when the discriminator can no longer distinguish the generator's output from the set of training examples.
We establish a theoretical results towards understanding this generator-discriminator training process.
arXiv Detail & Related papers (2020-03-09T10:50:35Z) - When Relation Networks meet GANs: Relation GANs with Triplet Loss [110.7572918636599]
Training stability is still a lingering concern of generative adversarial networks (GANs)
In this paper, we explore a relation network architecture for the discriminator and design a triplet loss which performs better generalization and stability.
Experiments on benchmark datasets show that the proposed relation discriminator and new loss can provide significant improvement on variable vision tasks.
arXiv Detail & Related papers (2020-02-24T11:35:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.