Towards Unsupervised Content Disentanglement in Sentence Representations
via Syntactic Roles
- URL: http://arxiv.org/abs/2206.11184v1
- Date: Wed, 22 Jun 2022 15:50:01 GMT
- Title: Towards Unsupervised Content Disentanglement in Sentence Representations
via Syntactic Roles
- Authors: Ghazi Felhi, Joseph Le Roux, Djam\'e Seddah
- Abstract summary: We develop an Attention-Driven Variational Autoencoder (ADVAE)
We show that it is possible to obtain representations of sentences where different syntactic roles correspond to clearly identified latent variables.
Our work constitutes a first step towards unsupervised controllable content generation.
- Score: 0.9582466286528458
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Linking neural representations to linguistic factors is crucial in order to
build and analyze NLP models interpretable by humans. Among these factors,
syntactic roles (e.g. subjects, direct objects,$\dots$) and their realizations
are essential markers since they can be understood as a decomposition of
predicative structures and thus the meaning of sentences. Starting from a deep
probabilistic generative model with attention, we measure the interaction
between latent variables and realizations of syntactic roles and show that it
is possible to obtain, without supervision, representations of sentences where
different syntactic roles correspond to clearly identified different latent
variables. The probabilistic model we propose is an Attention-Driven
Variational Autoencoder (ADVAE). Drawing inspiration from Transformer-based
machine translation models, ADVAEs enable the analysis of the interactions
between latent variables and input tokens through attention. We also develop an
evaluation protocol to measure disentanglement with regard to the realizations
of syntactic roles. This protocol is based on attention maxima for the encoder
and on latent variable perturbations for the decoder. Our experiments on raw
English text from the SNLI dataset show that $\textit{i)}$ disentanglement of
syntactic roles can be induced without supervision, $\textit{ii)}$ ADVAE
separates syntactic roles better than classical sequence VAEs and Transformer
VAEs, $\textit{iii)}$ realizations of syntactic roles can be separately
modified in sentences by mere intervention on the associated latent variables.
Our work constitutes a first step towards unsupervised controllable content
generation. The code for our work is publicly available.
Related papers
- Explaining Text Similarity in Transformer Models [52.571158418102584]
Recent advances in explainable AI have made it possible to mitigate limitations by leveraging improved explanations for Transformers.
We use BiLRP, an extension developed for computing second-order explanations in bilinear similarity models, to investigate which feature interactions drive similarity in NLP models.
Our findings contribute to a deeper understanding of different semantic similarity tasks and models, highlighting how novel explainable AI methods enable in-depth analyses and corpus-level insights.
arXiv Detail & Related papers (2024-05-10T17:11:31Z) - Probabilistic Transformer: A Probabilistic Dependency Model for
Contextual Word Representation [52.270712965271656]
We propose a new model of contextual word representation, not from a neural perspective, but from a purely syntactic and probabilistic perspective.
We find that the graph of our model resembles transformers, with correspondences between dependencies and self-attention.
Experiments show that our model performs competitively to transformers on small to medium sized datasets.
arXiv Detail & Related papers (2023-11-26T06:56:02Z) - Interpretable Sentence Representation with Variational Autoencoders and
Attention [0.685316573653194]
We develop methods to enhance the interpretability of recent representation learning techniques in natural language processing (NLP)
We leverage Variational Autoencoders (VAEs) due to their efficiency in relating observations to latent generative factors.
We build two models with inductive bias to separate information in latent representations into understandable concepts without annotated data.
arXiv Detail & Related papers (2023-05-04T13:16:15Z) - Disentangling Generative Factors in Natural Language with Discrete
Variational Autoencoders [0.0]
We argue that continuous variables may not be ideal to model features of textual data, due to the fact that most generative factors in text are discrete.
We propose a Variational Autoencoder based method which models language features as discrete variables and encourages independence between variables for learning disentangled representations.
arXiv Detail & Related papers (2021-09-15T09:10:05Z) - Prototypical Representation Learning for Relation Extraction [56.501332067073065]
This paper aims to learn predictive, interpretable, and robust relation representations from distantly-labeled data.
We learn prototypes for each relation from contextual information to best explore the intrinsic semantics of relations.
Results on several relation learning tasks show that our model significantly outperforms the previous state-of-the-art relational models.
arXiv Detail & Related papers (2021-03-22T08:11:43Z) - An Investigation of Language Model Interpretability via Sentence Editing [5.492504126672887]
We re-purpose a sentence editing dataset as a testbed for interpretability of pre-trained language models (PLMs)
This enables us to conduct a systematic investigation on an array of questions regarding PLMs' interpretability.
The investigation generates new insights, for example, contrary to the common understanding, we find that attention weights correlate well with human rationales.
arXiv Detail & Related papers (2020-11-28T00:46:43Z) - Disentangling Action Sequences: Discovering Correlated Samples [6.179793031975444]
We demonstrate the data itself plays a crucial role in disentanglement and instead of the factors, and the disentangled representations align the latent variables with the action sequences.
We propose a novel framework, fractional variational autoencoder (FVAE) to disentangle the action sequences with different significance step-by-step.
Experimental results on dSprites and 3D Chairs show that FVAE improves the stability of disentanglement.
arXiv Detail & Related papers (2020-10-17T07:37:50Z) - Is Supervised Syntactic Parsing Beneficial for Language Understanding?
An Empirical Investigation [71.70562795158625]
Traditional NLP has long held (supervised) syntactic parsing necessary for successful higher-level semantic language understanding (LU)
Recent advent of end-to-end neural models, self-supervised via language modeling (LM), and their success on a wide range of LU tasks, questions this belief.
We empirically investigate the usefulness of supervised parsing for semantic LU in the context of LM-pretrained transformer networks.
arXiv Detail & Related papers (2020-08-15T21:03:36Z) - Self-Attention Attribution: Interpreting Information Interactions Inside
Transformer [89.21584915290319]
We propose a self-attention attribution method to interpret the information interactions inside Transformer.
We show that the attribution results can be used as adversarial patterns to implement non-targeted attacks towards BERT.
arXiv Detail & Related papers (2020-04-23T14:58:22Z) - Improve Variational Autoencoder for Text Generationwith Discrete Latent
Bottleneck [52.08901549360262]
Variational autoencoders (VAEs) are essential tools in end-to-end representation learning.
VAEs tend to ignore latent variables with a strong auto-regressive decoder.
We propose a principled approach to enforce an implicit latent feature matching in a more compact latent space.
arXiv Detail & Related papers (2020-04-22T14:41:37Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.