Variational Hyper-Encoding Networks
- URL: http://arxiv.org/abs/2005.08482v2
- Date: Fri, 13 May 2022 00:20:28 GMT
- Title: Variational Hyper-Encoding Networks
- Authors: Phuoc Nguyen, Truyen Tran, Sunil Gupta, Santu Rana, Hieu-Chi Dam,
Svetha Venkatesh
- Abstract summary: We propose a framework called HyperVAE for encoding distributions of neural network parameters theta.
We predict the posterior distribution of the latent code, then use a matrix-network decoder to generate a posterior distribution q(theta)
- Score: 62.74164588885455
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose a framework called HyperVAE for encoding distributions of
distributions. When a target distribution is modeled by a VAE, its neural
network parameters \theta is drawn from a distribution p(\theta) which is
modeled by a hyper-level VAE. We propose a variational inference using Gaussian
mixture models to implicitly encode the parameters \theta into a low
dimensional Gaussian distribution. Given a target distribution, we predict the
posterior distribution of the latent code, then use a matrix-network decoder to
generate a posterior distribution q(\theta). HyperVAE can encode the parameters
\theta in full in contrast to common hyper-networks practices, which generate
only the scale and bias vectors as target-network parameters. Thus HyperVAE
preserves much more information about the model for each task in the latent
space. We discuss HyperVAE using the minimum description length (MDL) principle
and show that it helps HyperVAE to generalize. We evaluate HyperVAE in density
estimation tasks, outlier detection and discovery of novel design classes,
demonstrating its efficacy.
Related papers
Err
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.