Deep Learning Approaches to Lexical Simplification: A Survey
- URL: http://arxiv.org/abs/2305.12000v1
- Date: Fri, 19 May 2023 20:56:22 GMT
- Title: Deep Learning Approaches to Lexical Simplification: A Survey
- Authors: Kai North, Tharindu Ranasinghe, Matthew Shardlow, Marcos Zampieri
- Abstract summary: Lexical Simplification (LS) is the task of replacing complex for simpler words in a sentence.
LS is the lexical component of Text Simplification (TS)
Recent advances in deep learning have sparked renewed interest in LS.
- Score: 19.079916794185642
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Lexical Simplification (LS) is the task of replacing complex for simpler
words in a sentence whilst preserving the sentence's original meaning. LS is
the lexical component of Text Simplification (TS) with the aim of making texts
more accessible to various target populations. A past survey (Paetzold and
Specia, 2017) has provided a detailed overview of LS. Since this survey,
however, the AI/NLP community has been taken by storm by recent advances in
deep learning, particularly with the introduction of large language models
(LLM) and prompt learning. The high performance of these models sparked renewed
interest in LS. To reflect these recent advances, we present a comprehensive
survey of papers published between 2017 and 2023 on LS and its sub-tasks with a
special focus on deep learning. We also present benchmark datasets for the
future development of LS systems.
Related papers
- Scaling Up Summarization: Leveraging Large Language Models for Long Text Extractive Summarization [0.27624021966289597]
This paper introduces EYEGLAXS, a framework that leverages Large Language Models (LLMs) for extractive summarization.
EYEGLAXS focuses on extractive summarization to ensure factual and grammatical integrity.
The system sets new performance benchmarks on well-known datasets like PubMed and ArXiv.
arXiv Detail & Related papers (2024-08-28T13:52:19Z) - MultiLS: A Multi-task Lexical Simplification Framework [21.81108113189197]
We present MultiLS, the first LS framework that allows for the creation of a multi-task LS dataset.
We also present MultiLS-PT, the first dataset to be created using the MultiLS framework.
arXiv Detail & Related papers (2024-02-22T21:16:18Z) - An LLM-Enhanced Adversarial Editing System for Lexical Simplification [10.519804917399744]
Lexical Simplification aims to simplify text at the lexical level.
Existing methods rely heavily on annotated data.
We propose a novel LS method without parallel corpora.
arXiv Detail & Related papers (2024-02-22T17:04:30Z) - The Lay Person's Guide to Biomedicine: Orchestrating Large Language
Models [38.8292168447796]
Large language models (LLMs) have demonstrated a remarkable capacity for text simplification, background information generation, and text evaluation.
We propose a novel textitExplain-then-Summarise LS framework, which leverages LLMs to generate high-quality background knowledge.
We also propose two novel LS evaluation metrics, which assess layness from multiple perspectives.
arXiv Detail & Related papers (2024-02-21T03:21:14Z) - Large Language Models: A Survey [69.72787936480394]
Large Language Models (LLMs) have drawn a lot of attention due to their strong performance on a wide range of natural language tasks.
LLMs' ability of general-purpose language understanding and generation is acquired by training billions of model's parameters on massive amounts of text data.
arXiv Detail & Related papers (2024-02-09T05:37:09Z) - Rethinking Interpretability in the Era of Large Language Models [76.1947554386879]
Large language models (LLMs) have demonstrated remarkable capabilities across a wide array of tasks.
The capability to explain in natural language allows LLMs to expand the scale and complexity of patterns that can be given to a human.
These new capabilities raise new challenges, such as hallucinated explanations and immense computational costs.
arXiv Detail & Related papers (2024-01-30T17:38:54Z) - Supervised Knowledge Makes Large Language Models Better In-context Learners [94.89301696512776]
Large Language Models (LLMs) exhibit emerging in-context learning abilities through prompt engineering.
The challenge of improving the generalizability and factuality of LLMs in natural language understanding and question answering remains under-explored.
We propose a framework that enhances the reliability of LLMs as it: 1) generalizes out-of-distribution data, 2) elucidates how LLMs benefit from discriminative models, and 3) minimizes hallucinations in generative tasks.
arXiv Detail & Related papers (2023-12-26T07:24:46Z) - Evaluating, Understanding, and Improving Constrained Text Generation for Large Language Models [49.74036826946397]
This study investigates constrained text generation for large language models (LLMs)
Our research mainly focuses on mainstream open-source LLMs, categorizing constraints into lexical, structural, and relation-based types.
Results illuminate LLMs' capacity and deficiency to incorporate constraints and provide insights for future developments in constrained text generation.
arXiv Detail & Related papers (2023-10-25T03:58:49Z) - Harnessing Explanations: LLM-to-LM Interpreter for Enhanced
Text-Attributed Graph Representation Learning [51.90524745663737]
A key innovation is our use of explanations as features, which can be used to boost GNN performance on downstream tasks.
Our method achieves state-of-the-art results on well-established TAG datasets.
Our method significantly speeds up training, achieving a 2.88 times improvement over the closest baseline on ogbn-arxiv.
arXiv Detail & Related papers (2023-05-31T03:18:03Z) - ALEXSIS-PT: A New Resource for Portuguese Lexical Simplification [17.101023503289856]
ALEXSIS-PT is a novel multi-candidate dataset for Brazilian Portuguese LS containing 9,605 candidate substitutions for 387 complex words.
We evaluate four models for substitute generation on this dataset, namely mDistilBERT, mBERT, XLM-R, and BERTimbau.
arXiv Detail & Related papers (2022-09-19T14:10:21Z) - Shortcut Learning of Large Language Models in Natural Language
Understanding [119.45683008451698]
Large language models (LLMs) have achieved state-of-the-art performance on a series of natural language understanding tasks.
They might rely on dataset bias and artifacts as shortcuts for prediction.
This has significantly affected their generalizability and adversarial robustness.
arXiv Detail & Related papers (2022-08-25T03:51:39Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.