Gromov-Wasserstein Autoencoders
- URL: http://arxiv.org/abs/2209.07007v1
- Date: Thu, 15 Sep 2022 02:34:39 GMT
- Title: Gromov-Wasserstein Autoencoders
- Authors: Nao Nakagawa, Ren Togo, Takahiro Ogawa, Miki Haseyama
- Abstract summary: We propose a novel representation learning method, Gromov-Wasserstein Autoencoders (GWAE)
Instead of a likelihood-based objective, GWAE models have a trainable prior optimized by minimizing the Gromov-Wasserstein (GW) metric.
By restricting the family of the trainable prior, we can introduce meta-priors to control latent representations for downstream tasks.
- Score: 36.656435006076975
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Learning concise data representations without supervisory signals is a
fundamental challenge in machine learning. A prominent approach to this goal is
likelihood-based models such as variational autoencoders (VAE) to learn latent
representations based on a meta-prior, which is a general premise assumed
beneficial for downstream tasks (e.g., disentanglement). However, such
approaches often deviate from the original likelihood architecture to apply the
introduced meta-prior, causing undesirable changes in their training. In this
paper, we propose a novel representation learning method, Gromov-Wasserstein
Autoencoders (GWAE), which directly matches the latent and data distributions.
Instead of a likelihood-based objective, GWAE models have a trainable prior
optimized by minimizing the Gromov-Wasserstein (GW) metric. The GW metric
measures the distance structure-oriented discrepancy between distributions
supported on incomparable spaces, e.g., with different dimensionalities. By
restricting the family of the trainable prior, we can introduce meta-priors to
control latent representations for downstream tasks. The empirical comparison
with the existing VAE-based methods shows that GWAE models can learn
representations based on meta-priors by changing the prior family without
further modifying the GW objective.
Related papers
- CMVAE: Causal Meta VAE for Unsupervised Meta-Learning [3.0839245814393728]
Unsupervised meta-learning aims to learn the meta knowledge from unlabeled data and rapidly adapt to novel tasks.
Existing approaches may be misled by the context-bias from the training data.
We propose Causal Meta VAE (CMVAE) that encodes the priors into latent codes in the causal space and learns their relationships simultaneously to achieve the downstream few-shot image classification task.
arXiv Detail & Related papers (2023-02-20T02:49:35Z) - Outlier-Robust Gromov-Wasserstein for Graph Data [31.895380224961464]
We introduce a new and robust version of the Gromov-Wasserstein (GW) distance called RGW.
RGW features optimistically perturbed marginal constraints within a Kullback-Leibler divergence-based ambiguity set.
We demonstrate the effectiveness of RGW on real-world graph learning tasks, such as subgraph matching and partial shape correspondence.
arXiv Detail & Related papers (2023-02-09T12:57:29Z) - Self-Supervised Learning via Maximum Entropy Coding [57.56570417545023]
We propose Maximum Entropy Coding (MEC) as a principled objective that explicitly optimize on the structure of the representation.
MEC learns a more generalizable representation than previous methods based on specific pretext tasks.
It achieves state-of-the-art performance consistently on various downstream tasks, including not only ImageNet linear probe, but also semi-supervised classification, object detection, instance segmentation, and object tracking.
arXiv Detail & Related papers (2022-10-20T17:58:30Z) - Adaptive Convolutional Dictionary Network for CT Metal Artifact
Reduction [62.691996239590125]
We propose an adaptive convolutional dictionary network (ACDNet) for metal artifact reduction.
Our ACDNet can automatically learn the prior for artifact-free CT images via training data and adaptively adjust the representation kernels for each input CT image.
Our method inherits the clear interpretability of model-based methods and maintains the powerful representation ability of learning-based methods.
arXiv Detail & Related papers (2022-05-16T06:49:36Z) - A Sparsity-promoting Dictionary Model for Variational Autoencoders [16.61511959679188]
Structuring the latent space in deep generative models is important to yield more expressive models and interpretable representations.
We propose a simple yet effective methodology to structure the latent space via a sparsity-promoting dictionary model.
arXiv Detail & Related papers (2022-03-29T17:13:11Z) - Efficient Variational Graph Autoencoders for Unsupervised Cross-domain
Prerequisite Chains [3.358838755118655]
We introduce Domain-versaational Variational Graph Autoencoders (DAVGAE) to solve this cross-domain prerequisite chain learning task efficiently.
Our novel model consists of a variational graph autoencoder (VGAE) and a domain discriminator.
Results show that our model outperforms recent graph-based computation using only 1/10 graph scale and 1/3 time.
arXiv Detail & Related papers (2021-09-17T19:07:27Z) - Cauchy-Schwarz Regularized Autoencoder [68.80569889599434]
Variational autoencoders (VAE) are a powerful and widely-used class of generative models.
We introduce a new constrained objective based on the Cauchy-Schwarz divergence, which can be computed analytically for GMMs.
Our objective improves upon variational auto-encoding models in density estimation, unsupervised clustering, semi-supervised learning, and face analysis.
arXiv Detail & Related papers (2021-01-06T17:36:26Z) - Autoencoding Variational Autoencoder [56.05008520271406]
We study the implications of this behaviour on the learned representations and also the consequences of fixing it by introducing a notion of self consistency.
We show that encoders trained with our self-consistency approach lead to representations that are robust (insensitive) to perturbations in the input introduced by adversarial attacks.
arXiv Detail & Related papers (2020-12-07T14:16:14Z) - Variational Mutual Information Maximization Framework for VAE Latent
Codes with Continuous and Discrete Priors [5.317548969642376]
Variational Autoencoder (VAE) is a scalable method for learning directed latent variable models of complex data.
We propose Variational Mutual Information Maximization Framework for VAE to address this issue.
arXiv Detail & Related papers (2020-06-02T09:05:51Z) - Learning Meta Face Recognition in Unseen Domains [74.69681594452125]
We propose a novel face recognition method via meta-learning named Meta Face Recognition (MFR)
MFR synthesizes the source/target domain shift with a meta-optimization objective.
We propose two benchmarks for generalized face recognition evaluation.
arXiv Detail & Related papers (2020-03-17T14:10:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.