Hyperbolic VAE via Latent Gaussian Distributions
- URL: http://arxiv.org/abs/2209.15217v3
- Date: Sun, 29 Oct 2023 06:18:12 GMT
- Title: Hyperbolic VAE via Latent Gaussian Distributions
- Authors: Seunghyuk Cho, Juyong Lee, Dongwoo Kim
- Abstract summary: We propose a Gaussian manifold variational auto-encoder (GM-VAE) whose latent space consists of a set of Gaussian distributions.
In experiments, we demonstrate the efficacy of GM-VAE on two different tasks: density estimation of image datasets and environment modeling in model-based reinforcement learning.
- Score: 7.258394470200572
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a Gaussian manifold variational auto-encoder (GM-VAE) whose latent
space consists of a set of Gaussian distributions. It is known that the set of
the univariate Gaussian distributions with the Fisher information metric form a
hyperbolic space, which we call a Gaussian manifold. To learn the VAE endowed
with the Gaussian manifolds, we propose a pseudo-Gaussian manifold normal
distribution based on the Kullback-Leibler divergence, a local approximation of
the squared Fisher-Rao distance, to define a density over the latent space. In
experiments, we demonstrate the efficacy of GM-VAE on two different tasks:
density estimation of image datasets and environment modeling in model-based
reinforcement learning. GM-VAE outperforms the other variants of hyperbolic-
and Euclidean-VAEs on density estimation tasks and shows competitive
performance in model-based reinforcement learning. We observe that our model
provides strong numerical stability, addressing a common limitation reported in
previous hyperbolic-VAEs.
Related papers
- Ensemble Kalman filter in latent space using a variational autoencoder pair [0.2383122657918106]
variational autoencoder (VAE) is a machine learning (ML) technique that allows to map an arbitrary distribution to/from a latent space.
We propose a novel hybrid DA-ML approach in which VAEs are incorporated in the DA procedure.
arXiv Detail & Related papers (2025-02-18T16:11:05Z) - Theoretical Insights for Diffusion Guidance: A Case Study for Gaussian
Mixture Models [59.331993845831946]
Diffusion models benefit from instillation of task-specific information into the score function to steer the sample generation towards desired properties.
This paper provides the first theoretical study towards understanding the influence of guidance on diffusion models in the context of Gaussian mixture models.
arXiv Detail & Related papers (2024-03-03T23:15:48Z) - MGF: Mixed Gaussian Flow for Diverse Trajectory Prediction [72.70572835589158]
We propose constructing a mixed Gaussian prior for a normalizing flow model for trajectory prediction.
Our method achieves state-of-the-art performance in the evaluation of both trajectory alignment and diversity on the popular UCY/ETH and SDD datasets.
arXiv Detail & Related papers (2024-02-19T15:48:55Z) - Scaling Riemannian Diffusion Models [68.52820280448991]
We show that our method enables us to scale to high dimensional tasks on nontrivial manifold.
We model QCD densities on $SU(n)$ lattices and contrastively learned embeddings on high dimensional hyperspheres.
arXiv Detail & Related papers (2023-10-30T21:27:53Z) - Generalised Gaussian Process Latent Variable Models (GPLVM) with
Stochastic Variational Inference [9.468270453795409]
We study the doubly formulation of the BayesianVM model amenable with minibatch training.
We show how this framework is compatible with different latent variable formulations and perform experiments to compare a suite of models.
We demonstrate how we can train in the presence of massively missing data and obtain high-fidelity reconstructions.
arXiv Detail & Related papers (2022-02-25T21:21:51Z) - The Schr\"odinger Bridge between Gaussian Measures has a Closed Form [101.79851806388699]
We focus on the dynamic formulation of OT, also known as the Schr"odinger bridge (SB) problem.
In this paper, we provide closed-form expressions for SBs between Gaussian measures.
arXiv Detail & Related papers (2022-02-11T15:59:01Z) - Gaussian Graphical Models as an Ensemble Method for Distributed Gaussian
Processes [8.4159776055506]
We propose a novel approach for aggregating the Gaussian experts' predictions by Gaussian graphical model (GGM)
We first estimate the joint distribution of latent and observed variables using the Expectation-Maximization (EM) algorithm.
Our new method outperforms other state-of-the-art DGP approaches.
arXiv Detail & Related papers (2022-02-07T15:22:56Z) - Riemannian Score-Based Generative Modeling [56.20669989459281]
We introduce score-based generative models (SGMs) demonstrating remarkable empirical performance.
Current SGMs make the underlying assumption that the data is supported on a Euclidean manifold with flat geometry.
This prevents the use of these models for applications in robotics, geoscience or protein modeling.
arXiv Detail & Related papers (2022-02-06T11:57:39Z) - Variational Filtering with Copula Models for SLAM [5.242618356321224]
We show how it is possible to perform simultaneous localization and mapping (SLAM) with a larger class of distributions.
We integrate the distribution model with copulas into a Sequential Monte Carlo estimator and show how unknown model parameters can be learned through gradient-based optimization.
arXiv Detail & Related papers (2020-08-02T15:38:23Z) - Variational Hyper-Encoding Networks [62.74164588885455]
We propose a framework called HyperVAE for encoding distributions of neural network parameters theta.
We predict the posterior distribution of the latent code, then use a matrix-network decoder to generate a posterior distribution q(theta)
arXiv Detail & Related papers (2020-05-18T06:46:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.