Learning Manifold Dimensions with Conditional Variational Autoencoders
- URL: http://arxiv.org/abs/2302.11756v2
- Date: Tue, 13 Jun 2023 18:50:18 GMT
- Title: Learning Manifold Dimensions with Conditional Variational Autoencoders
- Authors: Yijia Zheng, Tong He, Yixuan Qiu, David Wipf
- Abstract summary: variational autoencoder (VAE) and its conditional extension (CVAE) are capable of state-of-the-art results across multiple domains.
We show that VAE global minima are indeed capable of recovering the correct manifold dimension.
We then extend this result to more general CVAEs, demonstrating practical scenarios.
- Score: 22.539599695796014
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Although the variational autoencoder (VAE) and its conditional extension
(CVAE) are capable of state-of-the-art results across multiple domains, their
precise behavior is still not fully understood, particularly in the context of
data (like images) that lie on or near a low-dimensional manifold. For example,
while prior work has suggested that the globally optimal VAE solution can learn
the correct manifold dimension, a necessary (but not sufficient) condition for
producing samples from the true data distribution, this has never been
rigorously proven. Moreover, it remains unclear how such considerations would
change when various types of conditioning variables are introduced, or when the
data support is extended to a union of manifolds (e.g., as is likely the case
for MNIST digits and related). In this work, we address these points by first
proving that VAE global minima are indeed capable of recovering the correct
manifold dimension. We then extend this result to more general CVAEs,
demonstrating practical scenarios whereby the conditioning variables allow the
model to adaptively learn manifolds of varying dimension across samples. Our
analyses, which have practical implications for various CVAE design choices,
are also supported by numerical results on both synthetic and real-world
datasets.
Related papers
- A Framework for Feasible Counterfactual Exploration incorporating Causality, Sparsity and Density [1.1782896991259]
The imminent need to interpret a Machine Learning model with counterfactual (CF) explanations has been notable in the research community.
This work uses different benchmark datasets to examine whether CF examples can be generated after a small amount of changes to the original input.
arXiv Detail & Related papers (2024-04-20T22:05:48Z) - VTAE: Variational Transformer Autoencoder with Manifolds Learning [144.0546653941249]
Deep generative models have demonstrated successful applications in learning non-linear data distributions through a number of latent variables.
The nonlinearity of the generator implies that the latent space shows an unsatisfactory projection of the data space, which results in poor representation learning.
We show that geodesics and accurate computation can substantially improve the performance of deep generative models.
arXiv Detail & Related papers (2023-04-03T13:13:19Z) - Convolutional Filtering on Sampled Manifolds [122.06927400759021]
We show that convolutional filtering on a sampled manifold converges to continuous manifold filtering.
Our findings are further demonstrated empirically on a problem of navigation control.
arXiv Detail & Related papers (2022-11-20T19:09:50Z) - Semi-Supervised Manifold Learning with Complexity Decoupled Chart
Autoencoders [65.2511270059236]
This work introduces a chart autoencoder with an asymmetric encoding-decoding process that can incorporate additional semi-supervised information such as class labels.
We discuss the theoretical approximation power of such networks that essentially depends on the intrinsic dimension of the data manifold.
arXiv Detail & Related papers (2022-08-22T19:58:03Z) - PAC Generalization via Invariant Representations [41.02828564338047]
We consider the notion of $epsilon$-approximate invariance in a finite sample setting.
Inspired by PAC learning, we obtain finite-sample out-of-distribution generalization guarantees.
Our results show bounds that do not scale in ambient dimension when intervention sites are restricted to lie in a constant size subset of in-degree bounded nodes.
arXiv Detail & Related papers (2022-05-30T15:50:14Z) - Measuring dissimilarity with diffeomorphism invariance [94.02751799024684]
We introduce DID, a pairwise dissimilarity measure applicable to a wide range of data spaces.
We prove that DID enjoys properties which make it relevant for theoretical study and practical use.
arXiv Detail & Related papers (2022-02-11T13:51:30Z) - Exploring Complementary Strengths of Invariant and Equivariant
Representations for Few-Shot Learning [96.75889543560497]
In many real-world problems, collecting a large number of labeled samples is infeasible.
Few-shot learning is the dominant approach to address this issue, where the objective is to quickly adapt to novel categories in presence of a limited number of samples.
We propose a novel training mechanism that simultaneously enforces equivariance and invariance to a general set of geometric transformations.
arXiv Detail & Related papers (2021-03-01T21:14:33Z) - Fork or Fail: Cycle-Consistent Training with Many-to-One Mappings [67.11712279612583]
Cycle-consistent training is widely used for learning a forward and inverse mapping between two domains of interest.
We develop a conditional variational autoencoder (CVAE) approach that can be viewed as converting surjective mappings to implicit bijections.
Our pipeline can capture such many-to-one mappings during cycle training while promoting graph-to-text diversity.
arXiv Detail & Related papers (2020-12-14T10:59:59Z) - Geometry-Aware Hamiltonian Variational Auto-Encoder [0.0]
Variational auto-encoders (VAEs) have proven to be a well suited tool for performing dimensionality reduction by extracting latent variables lying in a potentially much smaller dimensional space than the data.
However, such generative models may perform poorly when trained on small data sets which are abundant in many real-life fields such as medicine.
We argue that such latent space modelling provides useful information about its underlying structure leading to far more meaningfuls, more realistic data-generation and more reliable clustering.
arXiv Detail & Related papers (2020-10-22T08:26:46Z) - Variational Autoencoder with Learned Latent Structure [4.41370484305827]
We introduce the Variational Autoencoder with Learned Latent Structure (VAELLS)
VAELLS incorporates a learnable manifold model into the latent space of a VAE.
We validate our model on examples with known latent structure and also demonstrate its capabilities on a real-world dataset.
arXiv Detail & Related papers (2020-06-18T14:59:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.