A Statistical Analysis of Wasserstein Autoencoders for Intrinsically
Low-dimensional Data
- URL: http://arxiv.org/abs/2402.15710v1
- Date: Sat, 24 Feb 2024 04:13:40 GMT
- Title: A Statistical Analysis of Wasserstein Autoencoders for Intrinsically
Low-dimensional Data
- Authors: Saptarshi Chakraborty and Peter L. Bartlett
- Abstract summary: We show that Wasserstein Autoencoders (WAEs) can learn the data distributions when the network architectures are properly chosen.
We show that the convergence rates of the expected excess risk in the number of samples for WAEs are independent of the high feature dimension.
- Score: 38.964624328622
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Variational Autoencoders (VAEs) have gained significant popularity among
researchers as a powerful tool for understanding unknown distributions based on
limited samples. This popularity stems partly from their impressive performance
and partly from their ability to provide meaningful feature representations in
the latent space. Wasserstein Autoencoders (WAEs), a variant of VAEs, aim to
not only improve model efficiency but also interpretability. However, there has
been limited focus on analyzing their statistical guarantees. The matter is
further complicated by the fact that the data distributions to which WAEs are
applied - such as natural images - are often presumed to possess an underlying
low-dimensional structure within a high-dimensional feature space, which
current theory does not adequately account for, rendering known bounds
inefficient. To bridge the gap between the theory and practice of WAEs, in this
paper, we show that WAEs can learn the data distributions when the network
architectures are properly chosen. We show that the convergence rates of the
expected excess risk in the number of samples for WAEs are independent of the
high feature dimension, instead relying only on the intrinsic dimension of the
data distribution.
Related papers
- Adaptive Learning of the Latent Space of Wasserstein Generative Adversarial Networks [7.958528596692594]
We propose a novel framework called the latent Wasserstein GAN (LWGAN)
It fuses the Wasserstein auto-encoder and the Wasserstein GAN so that the intrinsic dimension of the data manifold can be adaptively learned.
We show that LWGAN is able to identify the correct intrinsic dimension under several scenarios.
arXiv Detail & Related papers (2024-09-27T01:25:22Z) - Concurrent Density Estimation with Wasserstein Autoencoders: Some
Statistical Insights [20.894503281724052]
Wasserstein Autoencoders (WAEs) have been a pioneering force in the realm of deep generative models.
Our work is an attempt to offer a theoretical understanding of the machinery behind WAEs.
arXiv Detail & Related papers (2023-12-11T18:27:25Z) - ProtoVAE: Prototypical Networks for Unsupervised Disentanglement [1.6114012813668934]
We introduce a novel deep generative VAE-based model, ProtoVAE, that leverages a deep metric learning Prototypical network trained using self-supervision.
Our model is completely unsupervised and requires no priori knowledge of the dataset, including the number of factors.
We evaluate our proposed model on the benchmark dSprites, 3DShapes, and MPI3D disentanglement datasets.
arXiv Detail & Related papers (2023-05-16T01:29:26Z) - RENs: Relevance Encoding Networks [0.0]
This paper proposes relevance encoding networks (RENs): a novel probabilistic VAE-based framework that uses the automatic relevance determination (ARD) prior in the latent space to learn the data-specific bottleneck dimensionality.
We show that the proposed model learns the relevant latent bottleneck dimensionality without compromising the representation and generation quality of the samples.
arXiv Detail & Related papers (2022-05-25T21:53:48Z) - Generative Counterfactuals for Neural Networks via Attribute-Informed
Perturbation [51.29486247405601]
We design a framework to generate counterfactuals for raw data instances with the proposed Attribute-Informed Perturbation (AIP)
By utilizing generative models conditioned with different attributes, counterfactuals with desired labels can be obtained effectively and efficiently.
Experimental results on real-world texts and images demonstrate the effectiveness, sample quality as well as efficiency of our designed framework.
arXiv Detail & Related papers (2021-01-18T08:37:13Z) - Efficient Robustness Certificates for Discrete Data: Sparsity-Aware
Randomized Smoothing for Graphs, Images and More [85.52940587312256]
We propose a model-agnostic certificate based on the randomized smoothing framework which subsumes earlier work and is tight, efficient, and sparsity-aware.
We show the effectiveness of our approach on a wide variety of models, datasets, and tasks -- specifically highlighting its use for Graph Neural Networks.
arXiv Detail & Related papers (2020-08-29T10:09:02Z) - Diversity inducing Information Bottleneck in Model Ensembles [73.80615604822435]
In this paper, we target the problem of generating effective ensembles of neural networks by encouraging diversity in prediction.
We explicitly optimize a diversity inducing adversarial loss for learning latent variables and thereby obtain diversity in the output predictions necessary for modeling multi-modal data.
Compared to the most competitive baselines, we show significant improvements in classification accuracy, under a shift in the data distribution.
arXiv Detail & Related papers (2020-03-10T03:10:41Z) - GANs with Conditional Independence Graphs: On Subadditivity of
Probability Divergences [70.30467057209405]
Generative Adversarial Networks (GANs) are modern methods to learn the underlying distribution of a data set.
GANs are designed in a model-free fashion where no additional information about the underlying distribution is available.
We propose a principled design of a model-based GAN that uses a set of simple discriminators on the neighborhoods of the Bayes-net/MRF.
arXiv Detail & Related papers (2020-03-02T04:31:22Z) - Distribution Approximation and Statistical Estimation Guarantees of
Generative Adversarial Networks [82.61546580149427]
Generative Adversarial Networks (GANs) have achieved a great success in unsupervised learning.
This paper provides approximation and statistical guarantees of GANs for the estimation of data distributions with densities in a H"older space.
arXiv Detail & Related papers (2020-02-10T16:47:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.