The Effects of Invertibility on the Representational Complexity of
Encoders in Variational Autoencoders
- URL: http://arxiv.org/abs/2107.04652v1
- Date: Fri, 9 Jul 2021 19:53:29 GMT
- Title: The Effects of Invertibility on the Representational Complexity of
Encoders in Variational Autoencoders
- Authors: Divyansh Pareek, Andrej Risteski
- Abstract summary: We show that if the generative map is "strongly invertible" (in a sense we suitably formalize), the inferential model need not be much more complex.
Importantly, we do not require the generative model to be layerwise invertible.
We provide theoretical support for the empirical wisdom that learning deep generative models is harder when data lies on a low-dimensional manifold.
- Score: 16.27499951949733
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Training and using modern neural-network based latent-variable generative
models (like Variational Autoencoders) often require simultaneously training a
generative direction along with an inferential(encoding) direction, which
approximates the posterior distribution over the latent variables. Thus, the
question arises: how complex does the inferential model need to be, in order to
be able to accurately model the posterior distribution of a given generative
model?
In this paper, we identify an important property of the generative map
impacting the required size of the encoder. We show that if the generative map
is "strongly invertible" (in a sense we suitably formalize), the inferential
model need not be much more complex. Conversely, we prove that there exist
non-invertible generative maps, for which the encoding direction needs to be
exponentially larger (under standard assumptions in computational complexity).
Importantly, we do not require the generative model to be layerwise invertible,
which a lot of the related literature assumes and isn't satisfied by many
architectures used in practice (e.g. convolution and pooling based networks).
Thus, we provide theoretical support for the empirical wisdom that learning
deep generative models is harder when data lies on a low-dimensional manifold.
Related papers
- Promises and Pitfalls of Generative Masked Language Modeling: Theoretical Framework and Practical Guidelines [74.42485647685272]
We focus on Generative Masked Language Models (GMLMs)
We train a model to fit conditional probabilities of the data distribution via masking, which are subsequently used as inputs to a Markov Chain to draw samples from the model.
We adapt the T5 model for iteratively-refined parallel decoding, achieving 2-3x speedup in machine translation with minimal sacrifice in quality.
arXiv Detail & Related papers (2024-07-22T18:00:00Z) - Disentanglement via Latent Quantization [60.37109712033694]
In this work, we construct an inductive bias towards encoding to and decoding from an organized latent space.
We demonstrate the broad applicability of this approach by adding it to both basic data-re (vanilla autoencoder) and latent-reconstructing (InfoGAN) generative models.
arXiv Detail & Related papers (2023-05-28T06:30:29Z) - VTAE: Variational Transformer Autoencoder with Manifolds Learning [144.0546653941249]
Deep generative models have demonstrated successful applications in learning non-linear data distributions through a number of latent variables.
The nonlinearity of the generator implies that the latent space shows an unsatisfactory projection of the data space, which results in poor representation learning.
We show that geodesics and accurate computation can substantially improve the performance of deep generative models.
arXiv Detail & Related papers (2023-04-03T13:13:19Z) - DIFFormer: Scalable (Graph) Transformers Induced by Energy Constrained
Diffusion [66.21290235237808]
We introduce an energy constrained diffusion model which encodes a batch of instances from a dataset into evolutionary states.
We provide rigorous theory that implies closed-form optimal estimates for the pairwise diffusion strength among arbitrary instance pairs.
Experiments highlight the wide applicability of our model as a general-purpose encoder backbone with superior performance in various tasks.
arXiv Detail & Related papers (2023-01-23T15:18:54Z) - Entropy optimized semi-supervised decomposed vector-quantized
variational autoencoder model based on transfer learning for multiclass text
classification and generation [3.9318191265352196]
We propose a semisupervised discrete latent variable model for multi-class text classification and text generation.
The proposed model employs the concept of transfer learning for training a quantized transformer model.
Experimental results indicate that the proposed model has surpassed the state-of-the-art models remarkably.
arXiv Detail & Related papers (2021-11-10T07:07:54Z) - Learning Disentangled Latent Factors from Paired Data in Cross-Modal
Retrieval: An Implicit Identifiable VAE Approach [33.61751393224223]
We deal with the problem of learning the underlying disentangled latent factors that are shared between the paired bi-modal data in cross-modal retrieval.
We propose a novel idea of the implicit decoder, which completely removes the ambient data decoding module from a latent variable model.
Our model is shown to identify the factors accurately, significantly outperforming conventional encoder-decoder latent variable models.
arXiv Detail & Related papers (2020-12-01T17:47:50Z) - Category-Learning with Context-Augmented Autoencoder [63.05016513788047]
Finding an interpretable non-redundant representation of real-world data is one of the key problems in Machine Learning.
We propose a novel method of using data augmentations when training autoencoders.
We train a Variational Autoencoder in such a way, that it makes transformation outcome predictable by auxiliary network.
arXiv Detail & Related papers (2020-10-10T14:04:44Z) - Goal-directed Generation of Discrete Structures with Conditional
Generative Models [85.51463588099556]
We introduce a novel approach to directly optimize a reinforcement learning objective, maximizing an expected reward.
We test our methodology on two tasks: generating molecules with user-defined properties and identifying short python expressions which evaluate to a given target value.
arXiv Detail & Related papers (2020-10-05T20:03:13Z) - When and How Can Deep Generative Models be Inverted? [28.83334026125828]
Deep generative models (GANs and VAEs) have been developed quite extensively in recent years.
We define conditions that are applicable to any inversion algorithm (gradient descent, deep encoder, etc.) under which such generative models are invertible.
We show that our method outperforms gradient descent when inverting such generators, both for clean and corrupted signals.
arXiv Detail & Related papers (2020-06-28T09:37:52Z) - Disentanglement by Nonlinear ICA with General Incompressible-flow
Networks (GIN) [30.74691299906988]
A central question of representation learning asks under which conditions it is possible to reconstruct the true latent variables of an arbitrarily complex generative process.
Recent breakthrough work by Khehem et al. on nonlinear ICA has answered this question for a broad class of conditional generative processes.
We extend this important result in a direction relevant for application to real-world data.
arXiv Detail & Related papers (2020-01-14T16:25:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.