Cognitive Simplification Operations Improve Text Simplification
- URL: http://arxiv.org/abs/2211.08825v1
- Date: Wed, 16 Nov 2022 10:51:03 GMT
- Title: Cognitive Simplification Operations Improve Text Simplification
- Authors: Eytan Chamovitz and Omri Abend
- Abstract summary: We present a method for incorporating knowledge from the cognitive accessibility domain into a Text Simplification model.
We show that by adding this inductive bias to a TS-trained model, it is able to adapt better to Cognitive Simplification without ever seeing CS data.
- Score: 24.970301040693883
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Text Simplification (TS) is the task of converting a text into a form that is
easier to read while maintaining the meaning of the original text. A sub-task
of TS is Cognitive Simplification (CS), converting text to a form that is
readily understood by people with cognitive disabilities without rendering it
childish or simplistic. This sub-task has yet to be explored with neural
methods in NLP, and resources for it are scarcely available. In this paper, we
present a method for incorporating knowledge from the cognitive accessibility
domain into a TS model, by introducing an inductive bias regarding what
simplification operations to use. We show that by adding this inductive bias to
a TS-trained model, it is able to adapt better to CS without ever seeing CS
data, and outperform a baseline model on a traditional TS benchmark. In
addition, we provide a novel test dataset for CS, and analyze the differences
between CS corpora and existing TS corpora, in terms of how simplification
operations are applied.
Related papers
- Analysing Zero-Shot Readability-Controlled Sentence Simplification [54.09069745799918]
We investigate how different types of contextual information affect a model's ability to generate sentences with the desired readability.
Results show that all tested models struggle to simplify sentences due to models' limitations and characteristics of the source sentences.
Our experiments also highlight the need for better automatic evaluation metrics tailored to RCTS.
arXiv Detail & Related papers (2024-09-30T12:36:25Z) - Injecting linguistic knowledge into BERT for Dialogue State Tracking [60.42231674887294]
This paper proposes a method that extracts linguistic knowledge via an unsupervised framework.
We then utilize this knowledge to augment BERT's performance and interpretability in Dialogue State Tracking (DST) tasks.
We benchmark this framework on various DST tasks and observe a notable improvement in accuracy.
arXiv Detail & Related papers (2023-11-27T08:38:42Z) - Learning Symbolic Rules over Abstract Meaning Representations for
Textual Reinforcement Learning [63.148199057487226]
We propose a modular, NEuroSymbolic Textual Agent (NESTA) that combines a generic semantic generalization with a rule induction system to learn interpretable rules as policies.
Our experiments show that the proposed NESTA method outperforms deep reinforcement learning-based techniques by achieving better to unseen test games and learning from fewer training interactions.
arXiv Detail & Related papers (2023-07-05T23:21:05Z) - On Conditional and Compositional Language Model Differentiable Prompting [75.76546041094436]
Prompts have been shown to be an effective method to adapt a frozen Pretrained Language Model (PLM) to perform well on downstream tasks.
We propose a new model, Prompt Production System (PRopS), which learns to transform task instructions or input metadata, into continuous prompts.
arXiv Detail & Related papers (2023-07-04T02:47:42Z) - Scalable Learning of Latent Language Structure With Logical Offline
Cycle Consistency [71.42261918225773]
Conceptually, LOCCO can be viewed as a form of self-learning where the semantic being trained is used to generate annotations for unlabeled text.
As an added bonus, the annotations produced by LOCCO can be trivially repurposed to train a neural text generation model.
arXiv Detail & Related papers (2023-05-31T16:47:20Z) - (Psycho-)Linguistic Features Meet Transformer Models for Improved
Explainable and Controllable Text Simplification [31.64341800095214]
We aim to advance current research on explainable and controllable TS.
We use a large set of (psycho-)linguistic features in combination with pre-trained language models to improve explainable complexity prediction.
We extend a state-of-the-art Seq2Seq TS model, ACCESS, to enable explicit control of ten attributes.
arXiv Detail & Related papers (2022-12-19T20:46:21Z) - Investigating Text Simplification Evaluation [21.128143745540292]
Modern text simplification (TS) heavily relies on the availability of gold standard data to build machine learning models.
Existing studies show that parallel TS corpora contain inaccurate simplifications and incorrect alignments.
evaluation is usually performed by using metrics such as BLEU or SARI to compare system output to the gold standard.
arXiv Detail & Related papers (2021-07-28T22:49:32Z) - A Survey on Text Simplification [0.0]
Text Simplification (TS) aims to reduce the linguistic complexity of content to make it easier to understand.
This survey seeks to provide a comprehensive overview of TS, including a brief description of earlier approaches used.
arXiv Detail & Related papers (2020-08-19T18:12:33Z) - Predicting What You Already Know Helps: Provable Self-Supervised
Learning [60.27658820909876]
Self-supervised representation learning solves auxiliary prediction tasks (known as pretext tasks) without requiring labeled data.
We show a mechanism exploiting the statistical connections between certain em reconstruction-based pretext tasks that guarantee to learn a good representation.
We prove the linear layer yields small approximation error even for complex ground truth function class.
arXiv Detail & Related papers (2020-08-03T17:56:13Z) - Deep Transformer based Data Augmentation with Subword Units for
Morphologically Rich Online ASR [0.0]
Deep Transformer models have proven to be particularly powerful in language modeling tasks for ASR.
Recent studies showed that a considerable part of the knowledge of neural network Language Models (LM) can be transferred to traditional n-grams by using neural text generation based data augmentation.
We show that although data augmentation with Transformer-generated text works well for isolating languages, it causes a vocabulary explosion in a morphologically rich language.
We propose a new method called subword-based neural text augmentation, where we retokenize the generated text into statistically derived subwords.
arXiv Detail & Related papers (2020-07-14T10:22:05Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.