Robust Training of Vector Quantized Bottleneck Models
- URL: http://arxiv.org/abs/2005.08520v1
- Date: Mon, 18 May 2020 08:23:41 GMT
- Title: Robust Training of Vector Quantized Bottleneck Models
- Authors: Adrian {\L}a\'ncucki, Jan Chorowski, Guillaume Sanchez, Ricard Marxer,
Nanxin Chen, Hans J.G.A. Dolfing, Sameer Khurana, Tanel Alum\"ae, Antoine
Laurent
- Abstract summary: We demonstrate methods for reliable and efficient training of discrete representation using Vector-Quantized Variational Auto-Encoder models (VQ-VAEs)
For unsupervised representation learning, they became viable alternatives to continuous latent variable models such as the Variational Auto-Encoder (VAE)
- Score: 21.540133031071438
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper we demonstrate methods for reliable and efficient training of
discrete representation using Vector-Quantized Variational Auto-Encoder models
(VQ-VAEs). Discrete latent variable models have been shown to learn nontrivial
representations of speech, applicable to unsupervised voice conversion and
reaching state-of-the-art performance on unit discovery tasks. For unsupervised
representation learning, they became viable alternatives to continuous latent
variable models such as the Variational Auto-Encoder (VAE). However, training
deep discrete variable models is challenging, due to the inherent
non-differentiability of the discretization operation. In this paper we focus
on VQ-VAE, a state-of-the-art discrete bottleneck model shown to perform on par
with its continuous counterparts. It quantizes encoder outputs with on-line
$k$-means clustering. We show that the codebook learning can suffer from poor
initialization and non-stationarity of clustered encoder outputs. We
demonstrate that these can be successfully overcome by increasing the learning
rate for the codebook and periodic date-dependent codeword re-initialization.
As a result, we achieve more robust training across different tasks, and
significantly increase the usage of latent codewords even for large codebooks.
This has practical benefit, for instance, in unsupervised representation
learning, where large codebooks may lead to disentanglement of latent
representations.
Related papers
- Vector Quantized Wasserstein Auto-Encoder [57.29764749855623]
We study learning deep discrete representations from the generative viewpoint.
We endow discrete distributions over sequences of codewords and learn a deterministic decoder that transports the distribution over the sequences of codewords to the data distribution.
We develop further theories to connect it with the clustering viewpoint of WS distance, allowing us to have a better and more controllable clustering solution.
arXiv Detail & Related papers (2023-02-12T13:51:36Z) - Discrete Key-Value Bottleneck [95.61236311369821]
Deep neural networks perform well on classification tasks where data streams are i.i.d. and labeled data is abundant.
One powerful approach that has addressed this challenge involves pre-training of large encoders on volumes of readily available data, followed by task-specific tuning.
Given a new task, however, updating the weights of these encoders is challenging as a large number of weights needs to be fine-tuned, and as a result, they forget information about the previous tasks.
We propose a model architecture to address this issue, building upon a discrete bottleneck containing pairs of separate and learnable key-value codes.
arXiv Detail & Related papers (2022-07-22T17:52:30Z) - Adaptive Discrete Communication Bottlenecks with Dynamic Vector
Quantization [76.68866368409216]
We propose learning to dynamically select discretization tightness conditioned on inputs.
We show that dynamically varying tightness in communication bottlenecks can improve model performance on visual reasoning and reinforcement learning tasks.
arXiv Detail & Related papers (2022-02-02T23:54:26Z) - Entropy optimized semi-supervised decomposed vector-quantized
variational autoencoder model based on transfer learning for multiclass text
classification and generation [3.9318191265352196]
We propose a semisupervised discrete latent variable model for multi-class text classification and text generation.
The proposed model employs the concept of transfer learning for training a quantized transformer model.
Experimental results indicate that the proposed model has surpassed the state-of-the-art models remarkably.
arXiv Detail & Related papers (2021-11-10T07:07:54Z) - Autoencoding Variational Autoencoder [56.05008520271406]
We study the implications of this behaviour on the learned representations and also the consequences of fixing it by introducing a notion of self consistency.
We show that encoders trained with our self-consistency approach lead to representations that are robust (insensitive) to perturbations in the input introduced by adversarial attacks.
arXiv Detail & Related papers (2020-12-07T14:16:14Z) - Unsupervised Controllable Generation with Self-Training [90.04287577605723]
controllable generation with GANs remains a challenging research problem.
We propose an unsupervised framework to learn a distribution of latent codes that control the generator through self-training.
Our framework exhibits better disentanglement compared to other variants such as the variational autoencoder.
arXiv Detail & Related papers (2020-07-17T21:50:35Z) - Improve Variational Autoencoder for Text Generationwith Discrete Latent
Bottleneck [52.08901549360262]
Variational autoencoders (VAEs) are essential tools in end-to-end representation learning.
VAEs tend to ignore latent variables with a strong auto-regressive decoder.
We propose a principled approach to enforce an implicit latent feature matching in a more compact latent space.
arXiv Detail & Related papers (2020-04-22T14:41:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.