Online Kernel based Generative Adversarial Networks
- URL: http://arxiv.org/abs/2006.11432v1
- Date: Fri, 19 Jun 2020 22:54:01 GMT
- Title: Online Kernel based Generative Adversarial Networks
- Authors: Yeojoon Youn, Neil Thistlethwaite, Sang Keun Choe, Jacob Abernethy
- Abstract summary: We show that Online Kernel-based Generative Adversarial Networks (OKGAN) mitigate a number of training issues, including mode collapse and cycling.
OKGANs empirically perform dramatically better, with respect to reverse KL-divergence, than other GAN formulations on synthetic data.
- Score: 0.45880283710344055
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: One of the major breakthroughs in deep learning over the past five years has
been the Generative Adversarial Network (GAN), a neural network-based
generative model which aims to mimic some underlying distribution given a
dataset of samples. In contrast to many supervised problems, where one tries to
minimize a simple objective function of the parameters, GAN training is
formulated as a min-max problem over a pair of network parameters. While
empirically GANs have shown impressive success in several domains, researchers
have been puzzled by unusual training behavior, including cycling so-called
mode collapse. In this paper, we begin by providing a quantitative method to
explore some of the challenges in GAN training, and we show empirically how
this relates fundamentally to the parametric nature of the discriminator
network. We propose a novel approach that resolves many of these issues by
relying on a kernel-based non-parametric discriminator that is highly amenable
to online training---we call this the Online Kernel-based Generative
Adversarial Networks (OKGAN). We show empirically that OKGANs mitigate a number
of training issues, including mode collapse and cycling, and are much more
amenable to theoretical guarantees. OKGANs empirically perform dramatically
better, with respect to reverse KL-divergence, than other GAN formulations on
synthetic data; on classical vision datasets such as MNIST, SVHN, and CelebA,
show comparable performance.
Related papers
- On the Convergence of (Stochastic) Gradient Descent for Kolmogorov--Arnold Networks [56.78271181959529]
Kolmogorov--Arnold Networks (KANs) have gained significant attention in the deep learning community.
Empirical investigations demonstrate that KANs optimized via gradient descent (SGD) are capable of achieving near-zero training loss.
arXiv Detail & Related papers (2024-10-10T15:34:10Z) - Generative Adversarial Reduced Order Modelling [0.0]
We present GAROM, a new approach for reduced order modelling (ROM) based on generative adversarial networks (GANs)
GANs have the potential to learn data distribution and generate more realistic data.
In this work, we combine the GAN and ROM framework, by introducing a data-driven generative adversarial model able to learn solutions to parametric differential equations.
arXiv Detail & Related papers (2023-05-25T09:23:33Z) - Generative Adversarial Method Based on Neural Tangent Kernels [13.664682865991255]
We propose a new generative algorithm called generative adversarial NTK (GA-NTK)
GA-NTK can generate images comparable to those by GANs but is much easier to train under various conditions.
We conduct extensive experiments on real-world datasets, and the results show that GA-NTK can generate images comparable to those by GANs but is much easier to train under various conditions.
arXiv Detail & Related papers (2022-04-08T14:17:46Z) - Self-Ensembling GAN for Cross-Domain Semantic Segmentation [107.27377745720243]
This paper proposes a self-ensembling generative adversarial network (SE-GAN) exploiting cross-domain data for semantic segmentation.
In SE-GAN, a teacher network and a student network constitute a self-ensembling model for generating semantic segmentation maps, which together with a discriminator, forms a GAN.
Despite its simplicity, we find SE-GAN can significantly boost the performance of adversarial training and enhance the stability of the model.
arXiv Detail & Related papers (2021-12-15T09:50:25Z) - Wide Network Learning with Differential Privacy [7.453881927237143]
Current generation of neural networks suffers significant loss accuracy under most practically relevant privacy training regimes.
We develop a general approach towards training these models that takes advantage of the sparsity of the gradients of private Empirical Minimization (ERM)
Following the same number of parameters, we propose a novel algorithm for privately training neural networks.
arXiv Detail & Related papers (2021-03-01T20:31:50Z) - Continual Learning in Recurrent Neural Networks [67.05499844830231]
We evaluate the effectiveness of continual learning methods for processing sequential data with recurrent neural networks (RNNs)
We shed light on the particularities that arise when applying weight-importance methods, such as elastic weight consolidation, to RNNs.
We show that the performance of weight-importance methods is not directly affected by the length of the processed sequences, but rather by high working memory requirements.
arXiv Detail & Related papers (2020-06-22T10:05:12Z) - Generative Adversarial Networks (GANs): An Overview of Theoretical
Model, Evaluation Metrics, and Recent Developments [9.023847175654602]
Generative Adversarial Network (GAN) is an effective method to produce samples of large-scale data distribution.
GANs provide an appropriate way to learn deep representations without widespread use of labeled training data.
In GANs, the generative model is estimated via a competitive process where the generator and discriminator networks are trained simultaneously.
arXiv Detail & Related papers (2020-05-27T05:56:53Z) - Feature Purification: How Adversarial Training Performs Robust Deep
Learning [66.05472746340142]
We show a principle that we call Feature Purification, where we show one of the causes of the existence of adversarial examples is the accumulation of certain small dense mixtures in the hidden weights during the training process of a neural network.
We present both experiments on the CIFAR-10 dataset to illustrate this principle, and a theoretical result proving that for certain natural classification tasks, training a two-layer neural network with ReLU activation using randomly gradient descent indeed this principle.
arXiv Detail & Related papers (2020-05-20T16:56:08Z) - When Relation Networks meet GANs: Relation GANs with Triplet Loss [110.7572918636599]
Training stability is still a lingering concern of generative adversarial networks (GANs)
In this paper, we explore a relation network architecture for the discriminator and design a triplet loss which performs better generalization and stability.
Experiments on benchmark datasets show that the proposed relation discriminator and new loss can provide significant improvement on variable vision tasks.
arXiv Detail & Related papers (2020-02-24T11:35:28Z) - Kernel and Rich Regimes in Overparametrized Models [69.40899443842443]
We show that gradient descent on overparametrized multilayer networks can induce rich implicit biases that are not RKHS norms.
We also demonstrate this transition empirically for more complex matrix factorization models and multilayer non-linear networks.
arXiv Detail & Related papers (2020-02-20T15:43:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.