Vector Quantized Wasserstein Auto-Encoder
- URL: http://arxiv.org/abs/2302.05917v2
- Date: Sat, 17 Jun 2023 06:52:21 GMT
- Title: Vector Quantized Wasserstein Auto-Encoder
- Authors: Tung-Long Vuong, Trung Le, He Zhao, Chuanxia Zheng, Mehrtash Harandi,
Jianfei Cai, Dinh Phung
- Abstract summary: We study learning deep discrete representations from the generative viewpoint.
We endow discrete distributions over sequences of codewords and learn a deterministic decoder that transports the distribution over the sequences of codewords to the data distribution.
We develop further theories to connect it with the clustering viewpoint of WS distance, allowing us to have a better and more controllable clustering solution.
- Score: 57.29764749855623
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: Learning deep discrete latent presentations offers a promise of better
symbolic and summarized abstractions that are more useful to subsequent
downstream tasks. Inspired by the seminal Vector Quantized Variational
Auto-Encoder (VQ-VAE), most of work in learning deep discrete representations
has mainly focused on improving the original VQ-VAE form and none of them has
studied learning deep discrete representations from the generative viewpoint.
In this work, we study learning deep discrete representations from the
generative viewpoint. Specifically, we endow discrete distributions over
sequences of codewords and learn a deterministic decoder that transports the
distribution over the sequences of codewords to the data distribution via
minimizing a WS distance between them. We develop further theories to connect
it with the clustering viewpoint of WS distance, allowing us to have a better
and more controllable clustering solution. Finally, we empirically evaluate our
method on several well-known benchmarks, where it achieves better qualitative
and quantitative performances than the other VQ-VAE variants in terms of the
codebook utilization and image reconstruction/generation.
Related papers
- Gaussian Mixture Vector Quantization with Aggregated Categorical Posterior [5.862123282894087]
We introduce the Vector Quantized Variational Autoencoder (VQ-VAE)
VQ-VAE is a type of variational autoencoder using discrete embedding as latent.
We show that GM-VQ improves codebook utilization and reduces information loss without relying on handcrafteds.
arXiv Detail & Related papers (2024-10-14T05:58:11Z) - Learning Invariant Molecular Representation in Latent Discrete Space [52.13724532622099]
We propose a new framework for learning molecular representations that exhibit invariance and robustness against distribution shifts.
Our model achieves stronger generalization against state-of-the-art baselines in the presence of various distribution shifts.
arXiv Detail & Related papers (2023-10-22T04:06:44Z) - Hierarchical Sketch Induction for Paraphrase Generation [79.87892048285819]
We introduce Hierarchical Refinement Quantized Variational Autoencoders (HRQ-VAE), a method for learning decompositions of dense encodings.
We use HRQ-VAE to encode the syntactic form of an input sentence as a path through the hierarchy, allowing us to more easily predict syntactic sketches at test time.
arXiv Detail & Related papers (2022-03-07T15:28:36Z) - Adaptive Discrete Communication Bottlenecks with Dynamic Vector
Quantization [76.68866368409216]
We propose learning to dynamically select discretization tightness conditioned on inputs.
We show that dynamically varying tightness in communication bottlenecks can improve model performance on visual reasoning and reinforcement learning tasks.
arXiv Detail & Related papers (2022-02-02T23:54:26Z) - Unsupervised Controllable Generation with Self-Training [90.04287577605723]
controllable generation with GANs remains a challenging research problem.
We propose an unsupervised framework to learn a distribution of latent codes that control the generator through self-training.
Our framework exhibits better disentanglement compared to other variants such as the variational autoencoder.
arXiv Detail & Related papers (2020-07-17T21:50:35Z) - Robust Training of Vector Quantized Bottleneck Models [21.540133031071438]
We demonstrate methods for reliable and efficient training of discrete representation using Vector-Quantized Variational Auto-Encoder models (VQ-VAEs)
For unsupervised representation learning, they became viable alternatives to continuous latent variable models such as the Variational Auto-Encoder (VAE)
arXiv Detail & Related papers (2020-05-18T08:23:41Z) - Improve Variational Autoencoder for Text Generationwith Discrete Latent
Bottleneck [52.08901549360262]
Variational autoencoders (VAEs) are essential tools in end-to-end representation learning.
VAEs tend to ignore latent variables with a strong auto-regressive decoder.
We propose a principled approach to enforce an implicit latent feature matching in a more compact latent space.
arXiv Detail & Related papers (2020-04-22T14:41:37Z) - Depthwise Discrete Representation Learning [2.728575246952532]
Recent advancements in learning Discrete Representations have led to state of art results in tasks that involve Language, Audio and Vision.
Some latent factors such as words, phonemes and shapes are better represented by discrete latent variables as opposed to continuous.
Vector Quantized Variational Autoencoders (VQVAE) have produced remarkable results in multiple domains.
arXiv Detail & Related papers (2020-04-11T18:57:13Z) - Deterministic Decoding for Discrete Data in Variational Autoencoders [5.254093731341154]
We study a VAE model with a deterministic decoder (DD-VAE) for sequential data that selects the highest-scoring tokens instead of sampling.
We demonstrate the performance of DD-VAE on multiple datasets, including molecular generation and optimization problems.
arXiv Detail & Related papers (2020-03-04T16:36:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.