Disentangling Generative Factors in Natural Language with Discrete
Variational Autoencoders
- URL: http://arxiv.org/abs/2109.07169v1
- Date: Wed, 15 Sep 2021 09:10:05 GMT
- Title: Disentangling Generative Factors in Natural Language with Discrete
Variational Autoencoders
- Authors: Giangiacomo Mercatali, Andr\'e Freitas
- Abstract summary: We argue that continuous variables may not be ideal to model features of textual data, due to the fact that most generative factors in text are discrete.
We propose a Variational Autoencoder based method which models language features as discrete variables and encourages independence between variables for learning disentangled representations.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The ability of learning disentangled representations represents a major step
for interpretable NLP systems as it allows latent linguistic features to be
controlled. Most approaches to disentanglement rely on continuous variables,
both for images and text. We argue that despite being suitable for image
datasets, continuous variables may not be ideal to model features of textual
data, due to the fact that most generative factors in text are discrete. We
propose a Variational Autoencoder based method which models language features
as discrete variables and encourages independence between variables for
learning disentangled representations. The proposed model outperforms
continuous and discrete baselines on several qualitative and quantitative
benchmarks for disentanglement as well as on a text style transfer downstream
application.
Related papers
- Probabilistic Transformer: A Probabilistic Dependency Model for
Contextual Word Representation [52.270712965271656]
We propose a new model of contextual word representation, not from a neural perspective, but from a purely syntactic and probabilistic perspective.
We find that the graph of our model resembles transformers, with correspondences between dependencies and self-attention.
Experiments show that our model performs competitively to transformers on small to medium sized datasets.
arXiv Detail & Related papers (2023-11-26T06:56:02Z) - Flow Factorized Representation Learning [109.51947536586677]
We introduce a generative model which specifies a distinct set of latent probability paths that define different input transformations.
We show that our model achieves higher likelihoods on standard representation learning benchmarks while simultaneously being closer to approximately equivariant models.
arXiv Detail & Related papers (2023-09-22T20:15:37Z) - Learning Semantic Textual Similarity via Topic-informed Discrete Latent
Variables [17.57873577962635]
We develop a topic-informed discrete latent variable model for semantic textual similarity.
Our model learns a shared latent space for sentence-pair representation via vector quantization.
We show that our model is able to surpass several strong neural baselines in semantic textual similarity tasks.
arXiv Detail & Related papers (2022-11-07T15:09:58Z) - Learning Disentangled Representations for Natural Language Definitions [0.0]
We argue that recurrent syntactic and semantic regularities in textual data can be used to provide the models with both structural biases and generative factors.
We leverage the semantic structures present in a representative and semantically dense category of sentence types, definitional sentences, for training a Variational Autoencoder to learn disentangled representations.
arXiv Detail & Related papers (2022-09-22T14:31:55Z) - Adaptive Discrete Communication Bottlenecks with Dynamic Vector
Quantization [76.68866368409216]
We propose learning to dynamically select discretization tightness conditioned on inputs.
We show that dynamically varying tightness in communication bottlenecks can improve model performance on visual reasoning and reinforcement learning tasks.
arXiv Detail & Related papers (2022-02-02T23:54:26Z) - Learning Conditional Invariance through Cycle Consistency [60.85059977904014]
We propose a novel approach to identify meaningful and independent factors of variation in a dataset.
Our method involves two separate latent subspaces for the target property and the remaining input information.
We demonstrate on synthetic and molecular data that our approach identifies more meaningful factors which lead to sparser and more interpretable models.
arXiv Detail & Related papers (2021-11-25T17:33:12Z) - Discrete Latent Variable Representations for Low-Resource Text
Classification [47.936293924113855]
We consider approaches to learning discrete latent variable models for text.
We compare the performance of the learned representations as features for low-resource document and sentence classification.
An amortized variant of Hard EM performs particularly well in the lowest-resource regimes.
arXiv Detail & Related papers (2020-06-11T06:55:13Z) - Improve Variational Autoencoder for Text Generationwith Discrete Latent
Bottleneck [52.08901549360262]
Variational autoencoders (VAEs) are essential tools in end-to-end representation learning.
VAEs tend to ignore latent variables with a strong auto-regressive decoder.
We propose a principled approach to enforce an implicit latent feature matching in a more compact latent space.
arXiv Detail & Related papers (2020-04-22T14:41:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.