GANs with Conditional Independence Graphs: On Subadditivity of
Probability Divergences
- URL: http://arxiv.org/abs/2003.00652v3
- Date: Thu, 25 Feb 2021 23:51:23 GMT
- Title: GANs with Conditional Independence Graphs: On Subadditivity of
Probability Divergences
- Authors: Mucong Ding, Constantinos Daskalakis, Soheil Feizi
- Abstract summary: Generative Adversarial Networks (GANs) are modern methods to learn the underlying distribution of a data set.
GANs are designed in a model-free fashion where no additional information about the underlying distribution is available.
We propose a principled design of a model-based GAN that uses a set of simple discriminators on the neighborhoods of the Bayes-net/MRF.
- Score: 70.30467057209405
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generative Adversarial Networks (GANs) are modern methods to learn the
underlying distribution of a data set. GANs have been widely used in sample
synthesis, de-noising, domain transfer, etc. GANs, however, are designed in a
model-free fashion where no additional information about the underlying
distribution is available. In many applications, however, practitioners have
access to the underlying independence graph of the variables, either as a
Bayesian network or a Markov Random Field (MRF). We ask: how can one use this
additional information in designing model-based GANs? In this paper, we provide
theoretical foundations to answer this question by studying subadditivity
properties of probability divergences, which establish upper bounds on the
distance between two high-dimensional distributions by the sum of distances
between their marginals over (local) neighborhoods of the graphical structure
of the Bayes-net or the MRF. We prove that several popular probability
divergences satisfy some notion of subadditivity under mild conditions. These
results lead to a principled design of a model-based GAN that uses a set of
simple discriminators on the neighborhoods of the Bayes-net/MRF, rather than a
giant discriminator on the entire network, providing significant statistical
and computational benefits. Our experiments on synthetic and real-world
datasets demonstrate the benefits of our principled design of model-based GANs.
Related papers
- Theory on Score-Mismatched Diffusion Models and Zero-Shot Conditional Samplers [49.97755400231656]
We present the first performance guarantee with explicit dimensional general score-mismatched diffusion samplers.
We show that score mismatches result in an distributional bias between the target and sampling distributions, proportional to the accumulated mismatch between the target and training distributions.
This result can be directly applied to zero-shot conditional samplers for any conditional model, irrespective of measurement noise.
arXiv Detail & Related papers (2024-10-17T16:42:12Z) - A Likelihood Based Approach to Distribution Regression Using Conditional Deep Generative Models [6.647819824559201]
We study the large-sample properties of a likelihood-based approach for estimating conditional deep generative models.
Our results lead to the convergence rate of a sieve maximum likelihood estimator for estimating the conditional distribution.
arXiv Detail & Related papers (2024-10-02T20:46:21Z) - Score Approximation, Estimation and Distribution Recovery of Diffusion
Models on Low-Dimensional Data [68.62134204367668]
This paper studies score approximation, estimation, and distribution recovery of diffusion models, when data are supported on an unknown low-dimensional linear subspace.
We show that with a properly chosen neural network architecture, the score function can be both accurately approximated and efficiently estimated.
The generated distribution based on the estimated score function captures the data geometric structures and converges to a close vicinity of the data distribution.
arXiv Detail & Related papers (2023-02-14T17:02:35Z) - Discovering Invariant Rationales for Graph Neural Networks [104.61908788639052]
Intrinsic interpretability of graph neural networks (GNNs) is to find a small subset of the input graph's features.
We propose a new strategy of discovering invariant rationale (DIR) to construct intrinsically interpretable GNNs.
arXiv Detail & Related papers (2022-01-30T16:43:40Z) - Non-Asymptotic Error Bounds for Bidirectional GANs [10.62911757343557]
We derive nearly sharp bounds for the bidirectional GAN (BiGAN) estimation error under the Dudley distance.
This is the first theoretical guarantee for the bidirectional GAN learning approach.
arXiv Detail & Related papers (2021-10-24T00:12:03Z) - An error analysis of generative adversarial networks for learning
distributions [11.842861158282265]
generative adversarial networks (GANs) learn probability distributions from finite samples.
GANs are able to adaptively learn data distributions with low-dimensional structure or have H"older densities.
Our analysis is based on a new oracle inequality decomposing the estimation error into generator and discriminator approximation error and statistical error.
arXiv Detail & Related papers (2021-05-27T08:55:19Z) - Generative Model without Prior Distribution Matching [26.91643368299913]
Variational Autoencoder (VAE) and its variations are classic generative models by learning a low-dimensional latent representation to satisfy some prior distribution.
We propose to let the prior match the embedding distribution rather than imposing the latent variables to fit the prior.
arXiv Detail & Related papers (2020-09-23T09:33:24Z) - Distribution Approximation and Statistical Estimation Guarantees of
Generative Adversarial Networks [82.61546580149427]
Generative Adversarial Networks (GANs) have achieved a great success in unsupervised learning.
This paper provides approximation and statistical guarantees of GANs for the estimation of data distributions with densities in a H"older space.
arXiv Detail & Related papers (2020-02-10T16:47:57Z) - Brainstorming Generative Adversarial Networks (BGANs): Towards
Multi-Agent Generative Models with Distributed Private Datasets [70.62568022925971]
generative adversarial networks (GANs) must be fed by large datasets that adequately represent the data space.
In many scenarios, the available datasets may be limited and distributed across multiple agents, each of which is seeking to learn the distribution of the data on its own.
In this paper, a novel brainstorming GAN (BGAN) architecture is proposed using which multiple agents can generate real-like data samples while operating in a fully distributed manner.
arXiv Detail & Related papers (2020-02-02T02:58:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.