Representation Of Lexical Stylistic Features In Language Models'
Embedding Space
- URL: http://arxiv.org/abs/2305.18657v2
- Date: Wed, 31 May 2023 22:50:25 GMT
- Title: Representation Of Lexical Stylistic Features In Language Models'
Embedding Space
- Authors: Qing Lyu, Marianna Apidianaki, Chris Callison-Burch
- Abstract summary: We show that it is possible to derive a vector representation for each of these stylistic notions from only a small number of seed pairs.
We conduct experiments on five datasets and find that static embeddings encode these features more accurately at the level of words and phrases.
The lower performance of contextualized representations at the word level is partially attributable to the anisotropy of their vector space.
- Score: 28.60690854046176
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: The representation space of pretrained Language Models (LMs) encodes rich
information about words and their relationships (e.g., similarity, hypernymy,
polysemy) as well as abstract semantic notions (e.g., intensity). In this
paper, we demonstrate that lexical stylistic notions such as complexity,
formality, and figurativeness, can also be identified in this space. We show
that it is possible to derive a vector representation for each of these
stylistic notions from only a small number of seed pairs. Using these vectors,
we can characterize new texts in terms of these dimensions by performing simple
calculations in the corresponding embedding space. We conduct experiments on
five datasets and find that static embeddings encode these features more
accurately at the level of words and phrases, whereas contextualized LMs
perform better on sentences. The lower performance of contextualized
representations at the word level is partially attributable to the anisotropy
of their vector space, which can be corrected to some extent using techniques
like standardization.
Related papers
- Unified Lexical Representation for Interpretable Visual-Language Alignment [52.059812317944434]
We introduce LexVLA, a more interpretable VLA framework by learning a unified lexical representation for both modalities without complex design.
We demonstrate that these two pre-trained uni-modal models can be well-aligned by fine-tuning on modest multi-modal dataset.
arXiv Detail & Related papers (2024-07-25T07:35:27Z) - Linear Spaces of Meanings: Compositional Structures in Vision-Language
Models [110.00434385712786]
We investigate compositional structures in data embeddings from pre-trained vision-language models (VLMs)
We first present a framework for understanding compositional structures from a geometric perspective.
We then explain what these structures entail probabilistically in the case of VLM embeddings, providing intuitions for why they arise in practice.
arXiv Detail & Related papers (2023-02-28T08:11:56Z) - Towards Robust and Semantically Organised Latent Representations for
Unsupervised Text Style Transfer [6.467090475885798]
We introduce EPAAEs (versading Perturbed Adrial AutoEncoders) which completes this perturbation model.
We empirically show that this (a) produces a better organised latent space that clusters stylistically similar sentences together.
We also extend the text style transfer tasks to NLI datasets and show that these more complex definitions of style are learned best by EPAAE.
arXiv Detail & Related papers (2022-05-04T20:04:24Z) - Word2Box: Learning Word Representation Using Box Embeddings [28.080105878687185]
Learning vector representations for words is one of the most fundamental topics in NLP.
Our model, Word2Box, takes a region-based approach to the problem of word representation, representing words as $n$-dimensional rectangles.
We demonstrate improved performance on various word similarity tasks, particularly on less common words.
arXiv Detail & Related papers (2021-06-28T01:17:11Z) - The Low-Dimensional Linear Geometry of Contextualized Word
Representations [27.50785941238007]
We study the linear geometry of contextualized word representations in ELMO and BERT.
We show that a variety of linguistic features are encoded in low-dimensional subspaces.
arXiv Detail & Related papers (2021-05-15T00:58:08Z) - Accurate Word Representations with Universal Visual Guidance [55.71425503859685]
This paper proposes a visual representation method to explicitly enhance conventional word embedding with multiple-aspect senses from visual guidance.
We build a small-scale word-image dictionary from a multimodal seed dataset where each word corresponds to diverse related images.
Experiments on 12 natural language understanding and machine translation tasks further verify the effectiveness and the generalization capability of the proposed approach.
arXiv Detail & Related papers (2020-12-30T09:11:50Z) - Unsupervised Distillation of Syntactic Information from Contextualized
Word Representations [62.230491683411536]
We tackle the task of unsupervised disentanglement between semantics and structure in neural language representations.
To this end, we automatically generate groups of sentences which are structurally similar but semantically different.
We demonstrate that our transformation clusters vectors in space by structural properties, rather than by lexical semantics.
arXiv Detail & Related papers (2020-10-11T15:13:18Z) - A Comparative Study on Structural and Semantic Properties of Sentence
Embeddings [77.34726150561087]
We propose a set of experiments using a widely-used large-scale data set for relation extraction.
We show that different embedding spaces have different degrees of strength for the structural and semantic properties.
These results provide useful information for developing embedding-based relation extraction methods.
arXiv Detail & Related papers (2020-09-23T15:45:32Z) - Learning Universal Representations from Word to Sentence [89.82415322763475]
This work introduces and explores the universal representation learning, i.e., embeddings of different levels of linguistic unit in a uniform vector space.
We present our approach of constructing analogy datasets in terms of words, phrases and sentences.
We empirically verify that well pre-trained Transformer models incorporated with appropriate training settings may effectively yield universal representation.
arXiv Detail & Related papers (2020-09-10T03:53:18Z) - Using Holographically Compressed Embeddings in Question Answering [0.0]
This research employs holographic compression of pre-trained embeddings to represent a token, its part-of-speech, and named entity type.
The implementation, in a modified question answering recurrent deep learning network, shows that semantic relationships are preserved, and yields strong performance.
arXiv Detail & Related papers (2020-07-14T18:29:49Z) - Word Equations: Inherently Interpretable Sparse Word Embeddingsthrough
Sparse Coding [0.0]
We create a system where each dimension is associated with some human understandable hint that can describe the meaning of that dimension.
We construct these embeddings through sparse coding, where each vector in the basis set is itself a word embedding.
We show that models trained using these sparse embeddings can achieve good performance and are more interpretable in practice, including through human evaluations.
arXiv Detail & Related papers (2020-04-08T19:49:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.