Adaptive Weighted Discriminator for Training Generative Adversarial
Networks
- URL: http://arxiv.org/abs/2012.03149v1
- Date: Sat, 5 Dec 2020 23:55:42 GMT
- Title: Adaptive Weighted Discriminator for Training Generative Adversarial
Networks
- Authors: Vasily Zadorozhnyy, Qiang Cheng, Qiang Ye
- Abstract summary: We introduce a new family of discriminator loss functions that adopts a weighted sum of real and fake parts.
Our method can be potentially applied to any discriminator model with a loss that is a sum of the real and fake parts.
- Score: 11.68198403603969
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generative adversarial network (GAN) has become one of the most important
neural network models for classical unsupervised machine learning. A variety of
discriminator loss functions have been developed to train GAN's discriminators
and they all have a common structure: a sum of real and fake losses that only
depends on the actual and generated data respectively. One challenge associated
with an equally weighted sum of two losses is that the training may benefit one
loss but harm the other, which we show causes instability and mode collapse. In
this paper, we introduce a new family of discriminator loss functions that
adopts a weighted sum of real and fake parts, which we call adaptive weighted
loss functions or aw-loss functions. Using the gradients of the real and fake
parts of the loss, we can adaptively choose weights to train a discriminator in
the direction that benefits the GAN's stability. Our method can be potentially
applied to any discriminator model with a loss that is a sum of the real and
fake parts. Experiments validated the effectiveness of our loss functions on an
unconditional image generation task, improving the baseline results by a
significant margin on CIFAR-10, STL-10, and CIFAR-100 datasets in Inception
Scores and FID.
Related papers
- Noise-Robust Loss Functions: Enhancing Bounded Losses for Large-Scale Noisy Data Learning [0.0]
Large annotated datasets inevitably contain noisy labels, which poses a major challenge for training deep neural networks as they easily memorize the labels.
Noise-robust loss functions have emerged as a notable strategy to counteract this issue, but it remains challenging to create a robust loss function which is not susceptible to underfitting.
We propose a novel method denoted as logit bias, which adds a real number $epsilon$ to the logit at the position of the correct class.
arXiv Detail & Related papers (2023-06-08T18:38:55Z) - Alternate Loss Functions for Classification and Robust Regression Can Improve the Accuracy of Artificial Neural Networks [6.452225158891343]
This paper shows that training speed and final accuracy of neural networks can significantly depend on the loss function used to train neural networks.
Two new classification loss functions that significantly improve performance on a wide variety of benchmark tasks are proposed.
arXiv Detail & Related papers (2023-03-17T12:52:06Z) - Xtreme Margin: A Tunable Loss Function for Binary Classification
Problems [0.0]
We provide an overview of a novel loss function, the Xtreme Margin loss function.
Unlike the binary cross-entropy and the hinge loss functions, this loss function provides researchers and practitioners flexibility with their training process.
arXiv Detail & Related papers (2022-10-31T22:39:32Z) - Improving GANs with A Dynamic Discriminator [106.54552336711997]
We argue that a discriminator with an on-the-fly adjustment on its capacity can better accommodate such a time-varying task.
A comprehensive empirical study confirms that the proposed training strategy, termed as DynamicD, improves the synthesis performance without incurring any additional cost or training objectives.
arXiv Detail & Related papers (2022-09-20T17:57:33Z) - Augmentation-Aware Self-Supervision for Data-Efficient GAN Training [68.81471633374393]
Training generative adversarial networks (GANs) with limited data is challenging because the discriminator is prone to overfitting.
We propose a novel augmentation-aware self-supervised discriminator that predicts the augmentation parameter of the augmented data.
We compare our method with state-of-the-art (SOTA) methods using the class-conditional BigGAN and unconditional StyleGAN2 architectures.
arXiv Detail & Related papers (2022-05-31T10:35:55Z) - Label Distributionally Robust Losses for Multi-class Classification:
Consistency, Robustness and Adaptivity [55.29408396918968]
We study a family of loss functions named label-distributionally robust (LDR) losses for multi-class classification.
Our contributions include both consistency and robustness by establishing top-$k$ consistency of LDR losses for multi-class classification.
We propose a new adaptive LDR loss that automatically adapts the individualized temperature parameter to the noise degree of class label of each instance.
arXiv Detail & Related papers (2021-12-30T00:27:30Z) - An Empirical Study on GANs with Margin Cosine Loss and Relativistic
Discriminator [4.899818550820575]
We introduce a new loss function, namely Relativistic Margin Cosine Loss (RMCosGAN)
We compare RMCosGAN performance with existing loss functions based on two metrics: Frechet inception distance and inception score.
The experimental results show that RMCosGAN outperforms the existing ones and significantly improves the quality of images generated.
arXiv Detail & Related papers (2021-10-21T17:25:47Z) - Center Prediction Loss for Re-identification [65.58923413172886]
We propose a new loss based on center predictivity, that is, a sample must be positioned in a location of the feature space such that from it we can roughly predict the location of the center of same-class samples.
We show that this new loss leads to a more flexible intra-class distribution constraint while ensuring the between-class samples are well-separated.
arXiv Detail & Related papers (2021-04-30T03:57:31Z) - $\sigma^2$R Loss: a Weighted Loss by Multiplicative Factors using
Sigmoidal Functions [0.9569316316728905]
We introduce a new loss function called squared reduction loss ($sigma2$R loss), which is regulated by a sigmoid function to inflate/deflate the error per instance.
Our loss has clear intuition and geometric interpretation, we demonstrate by experiments the effectiveness of our proposal.
arXiv Detail & Related papers (2020-09-18T12:34:40Z) - An Equivalence between Loss Functions and Non-Uniform Sampling in
Experience Replay [72.23433407017558]
We show that any loss function evaluated with non-uniformly sampled data can be transformed into another uniformly sampled loss function.
Surprisingly, we find in some environments PER can be replaced entirely by this new loss function without impact to empirical performance.
arXiv Detail & Related papers (2020-07-12T17:45:24Z) - When Relation Networks meet GANs: Relation GANs with Triplet Loss [110.7572918636599]
Training stability is still a lingering concern of generative adversarial networks (GANs)
In this paper, we explore a relation network architecture for the discriminator and design a triplet loss which performs better generalization and stability.
Experiments on benchmark datasets show that the proposed relation discriminator and new loss can provide significant improvement on variable vision tasks.
arXiv Detail & Related papers (2020-02-24T11:35:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.