Tessellated Wasserstein Auto-Encoders
- URL: http://arxiv.org/abs/2005.09923v2
- Date: Thu, 4 Mar 2021 02:08:40 GMT
- Title: Tessellated Wasserstein Auto-Encoders
- Authors: Kuo Gai and Shihua Zhang
- Abstract summary: Non-adversarial generative models such as variational auto-encoder (VAE), Wasserstein auto-encoders with maximum mean discrepancy (WAE-MMD), sliced-Wasserstein auto-encoder (SWAE) are relatively easy to train and have less mode collapse compared to Wasserstein auto-encoder with generative adversarial network (WAE-GAN)
We develop a novel framework called Tessellated Wasserstein Auto-encoders (TWAE) to tessellate the support of the target distribution into a given number of regions by the centroidal Voronoi t
- Score: 5.9394103049943485
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Non-adversarial generative models such as variational auto-encoder (VAE),
Wasserstein auto-encoders with maximum mean discrepancy (WAE-MMD),
sliced-Wasserstein auto-encoder (SWAE) are relatively easy to train and have
less mode collapse compared to Wasserstein auto-encoder with generative
adversarial network (WAE-GAN). However, they are not very accurate in
approximating the target distribution in the latent space because they don't
have a discriminator to detect the minor difference between real and fake. To
this end, we develop a novel non-adversarial framework called Tessellated
Wasserstein Auto-encoders (TWAE) to tessellate the support of the target
distribution into a given number of regions by the centroidal Voronoi
tessellation (CVT) technique and design batches of data according to the
tessellation instead of random shuffling for accurate computation of
discrepancy. Theoretically, we demonstrate that the error of estimate to the
discrepancy decreases when the numbers of samples $n$ and regions $m$ of the
tessellation become larger with rates of $\mathcal{O}(\frac{1}{\sqrt{n}})$ and
$\mathcal{O}(\frac{1}{\sqrt{m}})$, respectively. Given fixed $n$ and $m$, a
necessary condition for the upper bound of measurement error to be minimized is
that the tessellation is the one determined by CVT. TWAE is very flexible to
different non-adversarial metrics and can substantially enhance their
generative performance in terms of Fr\'{e}chet inception distance (FID)
compared to VAE, WAE-MMD, SWAE. Moreover, numerical results indeed demonstrate
that TWAE is competitive to the adversarial model WAE-GAN, demonstrating its
powerful generative ability.
Related papers
- Certifiably Robust Model Evaluation in Federated Learning under Meta-Distributional Shifts [8.700087812420687]
We provide guarantees for the model's performance on a different, unseen network "B"<n>We show how the principled vanilla DKW bound enables certification of the model's true performance on unseen clients within the same (source) network.
arXiv Detail & Related papers (2024-10-26T18:45:15Z) - A Statistical Analysis of Wasserstein Autoencoders for Intrinsically
Low-dimensional Data [38.964624328622]
We show that Wasserstein Autoencoders (WAEs) can learn the data distributions when the network architectures are properly chosen.
We show that the convergence rates of the expected excess risk in the number of samples for WAEs are independent of the high feature dimension.
arXiv Detail & Related papers (2024-02-24T04:13:40Z) - Concurrent Density Estimation with Wasserstein Autoencoders: Some
Statistical Insights [20.894503281724052]
Wasserstein Autoencoders (WAEs) have been a pioneering force in the realm of deep generative models.
Our work is an attempt to offer a theoretical understanding of the machinery behind WAEs.
arXiv Detail & Related papers (2023-12-11T18:27:25Z) - $t^3$-Variational Autoencoder: Learning Heavy-tailed Data with Student's
t and Power Divergence [7.0479532872043755]
$t3$VAE is a modified VAE framework that incorporates Student's t-distributions for the prior, encoder, and decoder.
We show that $t3$VAE significantly outperforms other models on CelebA and imbalanced CIFAR-100 datasets.
arXiv Detail & Related papers (2023-12-02T13:14:28Z) - Discrete Diffusion Modeling by Estimating the Ratios of the Data Distribution [67.9215891673174]
We propose score entropy as a novel loss that naturally extends score matching to discrete spaces.
We test our Score Entropy Discrete Diffusion models on standard language modeling tasks.
arXiv Detail & Related papers (2023-10-25T17:59:12Z) - DFedADMM: Dual Constraints Controlled Model Inconsistency for
Decentralized Federated Learning [52.83811558753284]
Decentralized learning (DFL) discards the central server and establishes a decentralized communication network.
Existing DFL methods still suffer from two major challenges: local inconsistency and local overfitting.
arXiv Detail & Related papers (2023-08-16T11:22:36Z) - Towards Faster Non-Asymptotic Convergence for Diffusion-Based Generative
Models [49.81937966106691]
We develop a suite of non-asymptotic theory towards understanding the data generation process of diffusion models.
In contrast to prior works, our theory is developed based on an elementary yet versatile non-asymptotic approach.
arXiv Detail & Related papers (2023-06-15T16:30:08Z) - FeDXL: Provable Federated Learning for Deep X-Risk Optimization [105.17383135458897]
We tackle a novel federated learning (FL) problem for optimizing a family of X-risks, to which no existing algorithms are applicable.
The challenges for designing an FL algorithm for X-risks lie in the non-decomability of the objective over multiple machines and the interdependency between different machines.
arXiv Detail & Related papers (2022-10-26T00:23:36Z) - Projected Sliced Wasserstein Autoencoder-based Hyperspectral Images
Anomaly Detection [42.585075865267946]
We propose the Projected Sliced Wasserstein (PSW) autoencoder-based anomaly detection method.
In particular, the computation-friendly eigen-decomposition method is leveraged to find the principal component for slicing the high-dimensional data.
Comprehensive experiments conducted on various real-world hyperspectral anomaly detection benchmarks demonstrate the superior performance of the proposed method.
arXiv Detail & Related papers (2021-12-20T09:21:02Z) - Unbiased Gradient Estimation for Variational Auto-Encoders using Coupled
Markov Chains [34.77971292478243]
The variational auto-encoder (VAE) is a deep latent variable model that has two neural networks in an autoencoder-like architecture.
We develop a training scheme for VAEs by introducing unbiased estimators of the log-likelihood gradient.
We show experimentally that VAEs fitted with unbiased estimators exhibit better predictive performance.
arXiv Detail & Related papers (2020-10-05T08:11:55Z) - Coded Stochastic ADMM for Decentralized Consensus Optimization with Edge
Computing [113.52575069030192]
Big data, including applications with high security requirements, are often collected and stored on multiple heterogeneous devices, such as mobile devices, drones and vehicles.
Due to the limitations of communication costs and security requirements, it is of paramount importance to extract information in a decentralized manner instead of aggregating data to a fusion center.
We consider the problem of learning model parameters in a multi-agent system with data locally processed via distributed edge nodes.
A class of mini-batch alternating direction method of multipliers (ADMM) algorithms is explored to develop the distributed learning model.
arXiv Detail & Related papers (2020-10-02T10:41:59Z) - Learning Autoencoders with Relational Regularization [89.53065887608088]
A new framework is proposed for learning autoencoders of data distributions.
We minimize the discrepancy between the model and target distributions, with a emphrelational regularization
We implement the framework with two scalable algorithms, making it applicable for both probabilistic and deterministic autoencoders.
arXiv Detail & Related papers (2020-02-07T17:27:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.