A New Paradigm for Generative Adversarial Networks based on Randomized
Decision Rules
- URL: http://arxiv.org/abs/2306.13641v1
- Date: Fri, 23 Jun 2023 17:50:34 GMT
- Title: A New Paradigm for Generative Adversarial Networks based on Randomized
Decision Rules
- Authors: Sehwan Kim, Qifan Song, and Faming Liang
- Abstract summary: The Generative Adversarial Network (GAN) was recently introduced in the literature as a novel machine learning method for training generative models.
It has many applications in statistics such as nonparametric clustering and nonparametric conditional independence tests.
In this paper, we identify the reasons why the GAN suffers from this issue, and to address it, we propose a new formulation for the GAN based on randomized decision rules.
- Score: 8.36840154574354
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Generative Adversarial Network (GAN) was recently introduced in the
literature as a novel machine learning method for training generative models.
It has many applications in statistics such as nonparametric clustering and
nonparametric conditional independence tests. However, training the GAN is
notoriously difficult due to the issue of mode collapse, which refers to the
lack of diversity among generated data. In this paper, we identify the reasons
why the GAN suffers from this issue, and to address it, we propose a new
formulation for the GAN based on randomized decision rules. In the new
formulation, the discriminator converges to a fixed point while the generator
converges to a distribution at the Nash equilibrium. We propose to train the
GAN by an empirical Bayes-like method by treating the discriminator as a
hyper-parameter of the posterior distribution of the generator. Specifically,
we simulate generators from its posterior distribution conditioned on the
discriminator using a stochastic gradient Markov chain Monte Carlo (MCMC)
algorithm, and update the discriminator using stochastic gradient descent along
with simulations of the generators. We establish convergence of the proposed
method to the Nash equilibrium. Apart from image generation, we apply the
proposed method to nonparametric clustering and nonparametric conditional
independence tests. A portion of the numerical results is presented in the
supplementary material.
Related papers
- Theory on Score-Mismatched Diffusion Models and Zero-Shot Conditional Samplers [49.97755400231656]
We present the first performance guarantee with explicit dimensional general score-mismatched diffusion samplers.
We show that score mismatches result in an distributional bias between the target and sampling distributions, proportional to the accumulated mismatch between the target and training distributions.
This result can be directly applied to zero-shot conditional samplers for any conditional model, irrespective of measurement noise.
arXiv Detail & Related papers (2024-10-17T16:42:12Z) - HJ-sampler: A Bayesian sampler for inverse problems of a stochastic process by leveraging Hamilton-Jacobi PDEs and score-based generative models [1.949927790632678]
This paper builds on the log transform known as the Cole-Hopf transform in Brownian motion contexts.
We develop a new algorithm, named the HJ-sampler, for inference for the inverse problem of a differential equation with given terminal observations.
arXiv Detail & Related papers (2024-09-15T05:30:54Z) - Wasserstein Generative Regression [6.366148239550708]
We propose a new and unified approach for nonparametric regression and conditional distribution learning.
Our approach simultaneously estimates a regression function and a conditional generator using a generative learning framework.
We provide theoretical guarantees by deriving non-asymptotic error bounds and the distributional consistency of our approach under suitable assumptions.
arXiv Detail & Related papers (2023-06-27T02:44:54Z) - Accurate generation of stochastic dynamics based on multi-model
Generative Adversarial Networks [0.0]
Generative Adversarial Networks (GANs) have shown immense potential in fields such as text and image generation.
Here we quantitatively test this approach by applying it to a prototypical process on a lattice.
Importantly, the discreteness of the model is retained despite the noise.
arXiv Detail & Related papers (2023-05-25T10:41:02Z) - Mode Penalty Generative Adversarial Network with adapted Auto-encoder [0.15229257192293197]
We propose a mode penalty GAN combined with pre-trained auto encoder for explicit representation of generated and real data samples in encoded space.
We demonstrate that applying the proposed method to GANs helps generator's optimization becoming more stable and having faster convergence through experimental evaluations.
arXiv Detail & Related papers (2020-11-16T03:39:53Z) - Identification of Probability weighted ARX models with arbitrary domains [75.91002178647165]
PieceWise Affine models guarantees universal approximation, local linearity and equivalence to other classes of hybrid system.
In this work, we focus on the identification of PieceWise Auto Regressive with eXogenous input models with arbitrary regions (NPWARX)
The architecture is conceived following the Mixture of Expert concept, developed within the machine learning field.
arXiv Detail & Related papers (2020-09-29T12:50:33Z) - GANs with Variational Entropy Regularizers: Applications in Mitigating
the Mode-Collapse Issue [95.23775347605923]
Building on the success of deep learning, Generative Adversarial Networks (GANs) provide a modern approach to learn a probability distribution from observed samples.
GANs often suffer from the mode collapse issue where the generator fails to capture all existing modes of the input distribution.
We take an information-theoretic approach and maximize a variational lower bound on the entropy of the generated samples to increase their diversity.
arXiv Detail & Related papers (2020-09-24T19:34:37Z) - Conditional Hybrid GAN for Sequence Generation [56.67961004064029]
We propose a novel conditional hybrid GAN (C-Hybrid-GAN) to solve this issue.
We exploit the Gumbel-Softmax technique to approximate the distribution of discrete-valued sequences.
We demonstrate that the proposed C-Hybrid-GAN outperforms the existing methods in context-conditioned discrete-valued sequence generation.
arXiv Detail & Related papers (2020-09-18T03:52:55Z) - Unsupervised Controllable Generation with Self-Training [90.04287577605723]
controllable generation with GANs remains a challenging research problem.
We propose an unsupervised framework to learn a distribution of latent codes that control the generator through self-training.
Our framework exhibits better disentanglement compared to other variants such as the variational autoencoder.
arXiv Detail & Related papers (2020-07-17T21:50:35Z) - The Bures Metric for Generative Adversarial Networks [10.69910379275607]
Generative Adversarial Networks (GANs) are performant generative methods yielding high-quality samples.
We propose to match the real batch diversity to the fake batch diversity.
We observe that diversity matching reduces mode collapse substantially and has a positive effect on the sample quality.
arXiv Detail & Related papers (2020-06-16T12:04:41Z) - Discriminator Contrastive Divergence: Semi-Amortized Generative Modeling
by Exploring Energy of the Discriminator [85.68825725223873]
Generative Adversarial Networks (GANs) have shown great promise in modeling high dimensional data.
We introduce the Discriminator Contrastive Divergence, which is well motivated by the property of WGAN's discriminator.
We demonstrate the benefits of significant improved generation on both synthetic data and several real-world image generation benchmarks.
arXiv Detail & Related papers (2020-04-05T01:50:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.