Improving Semantic Control in Discrete Latent Spaces with Transformer
Quantized Variational Autoencoders
- URL: http://arxiv.org/abs/2402.00723v1
- Date: Thu, 1 Feb 2024 16:14:35 GMT
- Title: Improving Semantic Control in Discrete Latent Spaces with Transformer
Quantized Variational Autoencoders
- Authors: Yingji Zhang, Danilo S. Carvalho, Marco Valentino, Ian Pratt-Hartmann,
Andre Freitas
- Abstract summary: We investigate discrete latent spaces in Vector Quantized Variational AutoEncoders (VQVAEs) to improve semantic control and generation in Transformer-based VAEs.
We propose T5VQVAE, a novel model that leverages the controllability of VQVAEs to guide the self-attention mechanism in T5 at the token-level.
Experimental results indicate that T5VQVAE outperforms existing state-of-the-art VAE models, including Optimus.
- Score: 5.037881619912574
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Achieving precise semantic control over the latent spaces of Variational
AutoEncoders (VAEs) holds significant value for downstream tasks in NLP as the
underlying generative mechanisms could be better localised, explained and
improved upon. Recent research, however, has struggled to achieve consistent
results, primarily due to the inevitable loss of semantic information in the
variational bottleneck and limited control over the decoding mechanism. To
overcome these challenges, we investigate discrete latent spaces in Vector
Quantized Variational AutoEncoders (VQVAEs) to improve semantic control and
generation in Transformer-based VAEs. In particular, We propose T5VQVAE, a
novel model that leverages the controllability of VQVAEs to guide the
self-attention mechanism in T5 at the token-level, exploiting its full
generalization capabilities. Experimental results indicate that T5VQVAE
outperforms existing state-of-the-art VAE models, including Optimus, in terms
of controllability and preservation of semantic information across different
tasks such as auto-encoding of sentences and mathematical expressions, text
transfer, and inference. Moreover, T5VQVAE exhibits improved inference
capabilities, suggesting potential applications for downstream natural language
and symbolic reasoning tasks.
Related papers
- Dynamic Context Adaptation and Information Flow Control in Transformers: Introducing the Evaluator Adjuster Unit and Gated Residual Connections [0.0]
This paper introduces two significant enhancements to the transformer architecture.
The Evaluator Unit (EAU) and Gated Residual Connections (GRC) are designed to address these limitations.
We evaluate the performance of these enhancements across several benchmarks in natural language processing.
arXiv Detail & Related papers (2024-05-22T07:33:24Z) - Exploring Latent Pathways: Enhancing the Interpretability of Autonomous Driving with a Variational Autoencoder [79.70947339175572]
A bio-inspired neural circuit policy model has emerged as an innovative control module.
We take a leap forward by integrating a variational autoencoder with the neural circuit policy controller.
In addition to the architectural shift toward a variational autoencoder, this study introduces the automatic latent perturbation tool.
arXiv Detail & Related papers (2024-04-02T09:05:47Z) - Graph-Induced Syntactic-Semantic Spaces in Transformer-Based Variational
AutoEncoders [5.037881619912574]
In this paper, we investigate latent space separation methods for structural syntactic injection in Transformer-based VAEs.
Specifically, we explore how syntactic structures can be leveraged in the encoding stage through the integration of graph-based and sequential models.
Our empirical evaluation, carried out on natural language sentences and mathematical expressions, reveals that the proposed end-to-end VAE architecture can result in a better overall organisation of the latent space.
arXiv Detail & Related papers (2023-11-14T22:47:23Z) - Optimizing Non-Autoregressive Transformers with Contrastive Learning [74.46714706658517]
Non-autoregressive Transformers (NATs) reduce the inference latency of Autoregressive Transformers (ATs) by predicting words all at once rather than in sequential order.
In this paper, we propose to ease the difficulty of modality learning via sampling from the model distribution instead of the data distribution.
arXiv Detail & Related papers (2023-05-23T04:20:13Z) - Recurrence Boosts Diversity! Revisiting Recurrent Latent Variable in
Transformer-Based Variational AutoEncoder for Diverse Text Generation [85.5379146125199]
Variational Auto-Encoder (VAE) has been widely adopted in text generation.
We propose TRACE, a Transformer-based recurrent VAE structure.
arXiv Detail & Related papers (2022-10-22T10:25:35Z) - Adaptive Discrete Communication Bottlenecks with Dynamic Vector
Quantization [76.68866368409216]
We propose learning to dynamically select discretization tightness conditioned on inputs.
We show that dynamically varying tightness in communication bottlenecks can improve model performance on visual reasoning and reinforcement learning tasks.
arXiv Detail & Related papers (2022-02-02T23:54:26Z) - Discrete Auto-regressive Variational Attention Models for Text Modeling [53.38382932162732]
Variational autoencoders (VAEs) have been widely applied for text modeling.
They are troubled by two challenges: information underrepresentation and posterior collapse.
We propose Discrete Auto-regressive Variational Attention Model (DAVAM) to address the challenges.
arXiv Detail & Related papers (2021-06-16T06:36:26Z) - Variational Autoencoder-Based Vehicle Trajectory Prediction with an
Interpretable Latent Space [0.0]
This paper introduces the Descriptive Variational Autoencoder (DVAE), an unsupervised and end-to-end trainable neural network for predicting vehicle trajectories.
The proposed model provides a similar prediction accuracy but with the great advantage of having an interpretable latent space.
arXiv Detail & Related papers (2021-03-25T10:15:53Z) - Transformer-based Conditional Variational Autoencoder for Controllable
Story Generation [39.577220559911055]
We investigate large-scale latent variable models (LVMs) for neural story generation with objectives in two threads: generation effectiveness and controllability.
We advocate to revive latent variable modeling, essentially the power of representation learning, in the era of Transformers.
Specifically, we integrate latent representation vectors with a Transformer-based pre-trained architecture to build conditional variational autoencoder (CVAE)
arXiv Detail & Related papers (2021-01-04T08:31:11Z) - Improve Variational Autoencoder for Text Generationwith Discrete Latent
Bottleneck [52.08901549360262]
Variational autoencoders (VAEs) are essential tools in end-to-end representation learning.
VAEs tend to ignore latent variables with a strong auto-regressive decoder.
We propose a principled approach to enforce an implicit latent feature matching in a more compact latent space.
arXiv Detail & Related papers (2020-04-22T14:41:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.