High-Dimensional Distribution Generation Through Deep Neural Networks
- URL: http://arxiv.org/abs/2107.12466v1
- Date: Mon, 26 Jul 2021 20:35:52 GMT
- Title: High-Dimensional Distribution Generation Through Deep Neural Networks
- Authors: Dmytro Perekrestenko, L\'eandre Eberhard, Helmut B\"olcskei
- Abstract summary: We show that every $d$-dimensional probability distribution of bounded support can be generated through deep ReLU networks.
We find that, for histogram target distributions, the number of bits needed to encode the corresponding generative network equals the fundamental limit for encoding probability distributions.
- Score: 2.141079906482723
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We show that every $d$-dimensional probability distribution of bounded
support can be generated through deep ReLU networks out of a $1$-dimensional
uniform input distribution. What is more, this is possible without incurring a
cost - in terms of approximation error measured in Wasserstein-distance -
relative to generating the $d$-dimensional target distribution from $d$
independent random variables. This is enabled by a vast generalization of the
space-filling approach discovered in (Bailey & Telgarsky, 2018). The
construction we propose elicits the importance of network depth in driving the
Wasserstein distance between the target distribution and its neural network
approximation to zero. Finally, we find that, for histogram target
distributions, the number of bits needed to encode the corresponding generative
network equals the fundamental limit for encoding probability distributions as
dictated by quantization theory.
Related papers
- Quantitative CLTs in Deep Neural Networks [12.845031126178593]
We study the distribution of a fully connected neural network with random Gaussian weights and biases.
We obtain quantitative bounds on normal approximations valid at large but finite $n$ and any fixed network depth.
Our bounds are strictly stronger in terms of their dependence on network width than any previously available in the literature.
arXiv Detail & Related papers (2023-07-12T11:35:37Z) - Robust Training and Verification of Implicit Neural Networks: A
Non-Euclidean Contractive Approach [64.23331120621118]
This paper proposes a theoretical and computational framework for training and robustness verification of implicit neural networks.
We introduce a related embedded network and show that the embedded network can be used to provide an $ell_infty$-norm box over-approximation of the reachable sets of the original network.
We apply our algorithms to train implicit neural networks on the MNIST dataset and compare the robustness of our models with the models trained via existing approaches in the literature.
arXiv Detail & Related papers (2022-08-08T03:13:24Z) - Learning Distributions by Generative Adversarial Networks: Approximation
and Generalization [0.6768558752130311]
We study how well generative adversarial networks learn from finite samples by analyzing the convergence rates of these models.
Our analysis is based on a new inequality oracle that decomposes the estimation error of GAN into the discriminator and generator approximation errors.
For generator approximation error, we show that neural network can approximately transform a low-dimensional source distribution to a high-dimensional target distribution.
arXiv Detail & Related papers (2022-05-25T09:26:17Z) - On the Effective Number of Linear Regions in Shallow Univariate ReLU
Networks: Convergence Guarantees and Implicit Bias [50.84569563188485]
We show that gradient flow converges in direction when labels are determined by the sign of a target network with $r$ neurons.
Our result may already hold for mild over- parameterization, where the width is $tildemathcalO(r)$ and independent of the sample size.
arXiv Detail & Related papers (2022-05-18T16:57:10Z) - Robust Estimation for Nonparametric Families via Generative Adversarial
Networks [92.64483100338724]
We provide a framework for designing Generative Adversarial Networks (GANs) to solve high dimensional robust statistics problems.
Our work extend these to robust mean estimation, second moment estimation, and robust linear regression.
In terms of techniques, our proposed GAN losses can be viewed as a smoothed and generalized Kolmogorov-Smirnov distance.
arXiv Detail & Related papers (2022-02-02T20:11:33Z) - On some theoretical limitations of Generative Adversarial Networks [77.34726150561087]
It is a general assumption that GANs can generate any probability distribution.
We provide a new result based on Extreme Value Theory showing that GANs can't generate heavy tailed distributions.
arXiv Detail & Related papers (2021-10-21T06:10:38Z) - On the capacity of deep generative networks for approximating
distributions [8.798333793391544]
We prove that neural networks can transform a one-dimensional source distribution to a distribution arbitrarily close to a high-dimensional target distribution in Wasserstein distances.
It is shown that the approximation error grows at most linearly on the ambient dimension.
$f$-divergences are less adequate than Waserstein distances as metrics of distributions for generating samples.
arXiv Detail & Related papers (2021-01-29T01:45:02Z) - Constructive Universal High-Dimensional Distribution Generation through
Deep ReLU Networks [3.4591414173342647]
We present an explicit deep neural network construction that transforms uniformly distributed one-dimensional noise into an arbitrarily close approximation of any two-dimensional Lipschitz-continuous target distribution.
We elicit the importance of depth - in our neural network construction - in driving the Wasserstein distance between the target distribution and the approximation realized by the network to zero.
arXiv Detail & Related papers (2020-06-30T10:36:15Z) - Variational Hyper-Encoding Networks [62.74164588885455]
We propose a framework called HyperVAE for encoding distributions of neural network parameters theta.
We predict the posterior distribution of the latent code, then use a matrix-network decoder to generate a posterior distribution q(theta)
arXiv Detail & Related papers (2020-05-18T06:46:09Z) - A Universal Approximation Theorem of Deep Neural Networks for Expressing
Probability Distributions [12.100913944042972]
We prove that there exists a deep neural network $g:mathbbRdrightarrow mathbbR$ with ReLU activation.
The size of neural network can grow exponentially in $d$ when $1$-Wasserstein distance is used as the discrepancy.
arXiv Detail & Related papers (2020-04-19T14:45:47Z) - Proving the Lottery Ticket Hypothesis: Pruning is All You Need [56.25432563818297]
The lottery ticket hypothesis states that a randomly-d network contains a small subnetwork such that, when trained in isolation, can compete with the performance of the original network.
We prove an even stronger hypothesis, showing that for every bounded distribution and every target network with bounded weights, a sufficiently over- parameterized neural network with random weights contains a subnetwork with roughly the same accuracy as the target network, without any further training.
arXiv Detail & Related papers (2020-02-03T07:23:11Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.