Lexical Substitution is not Synonym Substitution: On the Importance of Producing Contextually Relevant Word Substitutes
- URL: http://arxiv.org/abs/2502.04173v1
- Date: Thu, 06 Feb 2025 16:05:50 GMT
- Title: Lexical Substitution is not Synonym Substitution: On the Importance of Producing Contextually Relevant Word Substitutes
- Authors: Juraj Vladika, Stephen Meisenbacher, Florian Matthes,
- Abstract summary: We introduce ConCat, a simple augmented approach which utilizes the original sentence to bolster contextual information sent to the model.
Our study includes a quantitative evaluation, measured via sentence similarity and task performance.
We also conduct a qualitative human analysis to validate that users prefer the substitutions proposed by our method, as opposed to previous methods.
- Score: 5.065947993017158
- License:
- Abstract: Lexical Substitution is the task of replacing a single word in a sentence with a similar one. This should ideally be one that is not necessarily only synonymous, but also fits well into the surrounding context of the target word, while preserving the sentence's grammatical structure. Recent advances in Lexical Substitution have leveraged the masked token prediction task of Pre-trained Language Models to generate replacements for a given word in a sentence. With this technique, we introduce ConCat, a simple augmented approach which utilizes the original sentence to bolster contextual information sent to the model. Compared to existing approaches, it proves to be very effective in guiding the model to make contextually relevant predictions for the target word. Our study includes a quantitative evaluation, measured via sentence similarity and task performance. In addition, we conduct a qualitative human analysis to validate that users prefer the substitutions proposed by our method, as opposed to previous methods. Finally, we test our approach on the prevailing benchmark for Lexical Substitution, CoInCo, revealing potential pitfalls of the benchmark. These insights serve as the foundation for a critical discussion on the way in which Lexical Substitution is evaluated.
Related papers
- Learning to Substitute Words with Model-based Score Ranking [17.46170699645023]
Smart word substitution aims to enhance sentence quality by improving word choices.
Current benchmarks rely on human-labeled data.
We employ a model-based score (BARTScore) to quantify sentence quality.
arXiv Detail & Related papers (2025-02-09T15:26:32Z) - Analyzing Semantic Change through Lexical Replacements [2.509907053583601]
We study the effect of unexpected contexts introduced by textitlexical replacements
We propose a textitreplacement schema where a target word is substituted with lexical replacements of varying relatedness.
We are the first to evaluate the use of LLaMa for semantic change detection.
arXiv Detail & Related papers (2024-04-29T10:20:41Z) - ParaLS: Lexical Substitution via Pretrained Paraphraser [18.929859707202517]
This study explores how to generate the substitute candidates from a paraphraser.
We propose two simple decoding strategies that focus on the variations of the target word during decoding.
arXiv Detail & Related papers (2023-05-14T12:49:16Z) - Textual Entailment Recognition with Semantic Features from Empirical
Text Representation [60.31047947815282]
A text entails a hypothesis if and only if the true value of the hypothesis follows the text.
In this paper, we propose a novel approach to identifying the textual entailment relationship between text and hypothesis.
We employ an element-wise Manhattan distance vector-based feature that can identify the semantic entailment relationship between the text-hypothesis pair.
arXiv Detail & Related papers (2022-10-18T10:03:51Z) - Unsupervised Lexical Substitution with Decontextualised Embeddings [48.00929769805882]
We propose a new unsupervised method for lexical substitution using pre-trained language models.
Our method retrieves substitutes based on the similarity of contextualised and decontextualised word embeddings.
We conduct experiments in English and Italian, and show that our method substantially outperforms strong baselines.
arXiv Detail & Related papers (2022-09-17T03:51:47Z) - Semantic-Preserving Adversarial Text Attacks [85.32186121859321]
We propose a Bigram and Unigram based adaptive Semantic Preservation Optimization (BU-SPO) method to examine the vulnerability of deep models.
Our method achieves the highest attack success rates and semantics rates by changing the smallest number of words compared with existing methods.
arXiv Detail & Related papers (2021-08-23T09:05:18Z) - LexSubCon: Integrating Knowledge from Lexical Resources into Contextual
Embeddings for Lexical Substitution [76.615287796753]
We introduce LexSubCon, an end-to-end lexical substitution framework based on contextual embedding models.
This is achieved by combining contextual information with knowledge from structured lexical resources.
Our experiments show that LexSubCon outperforms previous state-of-the-art methods on LS07 and CoInCo benchmark datasets.
arXiv Detail & Related papers (2021-07-11T21:25:56Z) - Swords: A Benchmark for Lexical Substitution with Improved Data Coverage
and Quality [126.55416118361495]
We release a new benchmark for lexical substitution, the task of finding appropriate substitutes for a target word in a context.
We use a context-free thesaurus to produce candidates and rely on human judgement to determine contextual appropriateness.
Compared to the previous largest benchmark, our Swords benchmark has 4.1x more substitutes per target word for the same level of quality, and its substitutes are 1.5x more appropriate (based on human judgement) for the same number of substitutes.
arXiv Detail & Related papers (2021-06-08T04:58:29Z) - Neural Syntactic Preordering for Controlled Paraphrase Generation [57.5316011554622]
Our work uses syntactic transformations to softly "reorder'' the source sentence and guide our neural paraphrasing model.
First, given an input sentence, we derive a set of feasible syntactic rearrangements using an encoder-decoder model.
Next, we use each proposed rearrangement to produce a sequence of position embeddings, which encourages our final encoder-decoder paraphrase model to attend to the source words in a particular order.
arXiv Detail & Related papers (2020-05-05T09:02:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.