A Unified Approach to Variational Autoencoders and Stochastic
Normalizing Flows via Markov Chains
- URL: http://arxiv.org/abs/2111.12506v1
- Date: Wed, 24 Nov 2021 14:04:32 GMT
- Title: A Unified Approach to Variational Autoencoders and Stochastic
Normalizing Flows via Markov Chains
- Authors: Johannes Hertrich, Paul Hagemann, Gabriele Steidl
- Abstract summary: We provide a unified framework to handle normalizing flows and variational autoencoders via Markov chains.
Our framework establishes a useful mathematical tool to combine the various approaches.
- Score: 0.45119235878273
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Normalizing flows, diffusion normalizing flows and variational autoencoders
are powerful generative models. In this paper, we provide a unified framework
to handle these approaches via Markov chains. Indeed, we consider stochastic
normalizing flows as pair of Markov chains fulfilling some properties and show
that many state-of-the-art models for data generation fit into this framework.
The Markov chains point of view enables us to couple both deterministic layers
as invertible neural networks and stochastic layers as Metropolis-Hasting
layers, Langevin layers and variational autoencoders in a mathematically sound
way. Besides layers with densities as Langevin layers, diffusion layers or
variational autoencoders, also layers having no densities as deterministic
layers or Metropolis-Hasting layers can be handled. Hence our framework
establishes a useful mathematical tool to combine the various approaches.
Related papers
- Mechanistic Permutability: Match Features Across Layers [4.2056926734482065]
We introduce SAE Match, a novel, data-free method for aligning SAE features across different layers of a neural network.
Our work advances the understanding of feature dynamics in neural networks and provides a new tool for mechanistic interpretability studies.
arXiv Detail & Related papers (2024-10-10T06:55:38Z) - Variational Classification [51.2541371924591]
We derive a variational objective to train the model, analogous to the evidence lower bound (ELBO) used to train variational auto-encoders.
Treating inputs to the softmax layer as samples of a latent variable, our abstracted perspective reveals a potential inconsistency.
We induce a chosen latent distribution, instead of the implicit assumption found in a standard softmax layer.
arXiv Detail & Related papers (2023-05-17T17:47:19Z) - Multiscale Graph Neural Network Autoencoders for Interpretable
Scientific Machine Learning [0.0]
The goal of this work is to address two limitations in autoencoder-based models: latent space interpretability and compatibility with unstructured meshes.
This is accomplished here with the development of a novel graph neural network (GNN) autoencoding architecture with demonstrations on complex fluid flow applications.
arXiv Detail & Related papers (2023-02-13T08:47:11Z) - Modiff: Action-Conditioned 3D Motion Generation with Denoising Diffusion
Probabilistic Models [58.357180353368896]
We propose a conditional paradigm that benefits from the denoising diffusion probabilistic model (DDPM) to tackle the problem of realistic and diverse action-conditioned 3D skeleton-based motion generation.
We are a pioneering attempt that uses DDPM to synthesize a variable number of motion sequences conditioned on a categorical action.
arXiv Detail & Related papers (2023-01-10T13:15:42Z) - WLD-Reg: A Data-dependent Within-layer Diversity Regularizer [98.78384185493624]
Neural networks are composed of multiple layers arranged in a hierarchical structure jointly trained with a gradient-based optimization.
We propose to complement this traditional 'between-layer' feedback with additional 'within-layer' feedback to encourage the diversity of the activations within the same layer.
We present an extensive empirical study confirming that the proposed approach enhances the performance of several state-of-the-art neural network models in multiple tasks.
arXiv Detail & Related papers (2023-01-03T20:57:22Z) - A new perspective on probabilistic image modeling [92.89846887298852]
We present a new probabilistic approach for image modeling capable of density estimation, sampling and tractable inference.
DCGMMs can be trained end-to-end by SGD from random initial conditions, much like CNNs.
We show that DCGMMs compare favorably to several recent PC and SPN models in terms of inference, classification and sampling.
arXiv Detail & Related papers (2022-03-21T14:53:57Z) - Mesh Convolutional Autoencoder for Semi-Regular Meshes of Different
Sizes [0.0]
State-of-the-art mesh convolutional autoencoders require a fixed connectivity of all input meshes handled by the autoencoder.
We transform the discretization of the surfaces to semi-regular meshes that have a locally regular connectivity and whose meshing is hierarchical.
We apply the same mesh autoencoder to different datasets and our reconstruction error is more than 50% lower than the error from state-of-the-art models.
arXiv Detail & Related papers (2021-10-18T15:30:40Z) - Deep Autoencoding Topic Model with Scalable Hybrid Bayesian Inference [55.35176938713946]
We develop deep autoencoding topic model (DATM) that uses a hierarchy of gamma distributions to construct its multi-stochastic-layer generative network.
We propose a Weibull upward-downward variational encoder that deterministically propagates information upward via a deep neural network, followed by a downward generative model.
The efficacy and scalability of our models are demonstrated on both unsupervised and supervised learning tasks on big corpora.
arXiv Detail & Related papers (2020-06-15T22:22:56Z) - DHP: Differentiable Meta Pruning via HyperNetworks [158.69345612783198]
This paper introduces a differentiable pruning method via hypernetworks for automatic network pruning.
Latent vectors control the output channels of the convolutional layers in the backbone network and act as a handle for the pruning of the layers.
Experiments are conducted on various networks for image classification, single image super-resolution, and denoising.
arXiv Detail & Related papers (2020-03-30T17:59:18Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.