(Psycho-)Linguistic Features Meet Transformer Models for Improved
Explainable and Controllable Text Simplification
- URL: http://arxiv.org/abs/2212.09848v1
- Date: Mon, 19 Dec 2022 20:46:21 GMT
- Title: (Psycho-)Linguistic Features Meet Transformer Models for Improved
Explainable and Controllable Text Simplification
- Authors: Yu Qiao, Xiaofei Li, Daniel Wiechmann, Elma Kerz
- Abstract summary: We aim to advance current research on explainable and controllable TS.
We use a large set of (psycho-)linguistic features in combination with pre-trained language models to improve explainable complexity prediction.
We extend a state-of-the-art Seq2Seq TS model, ACCESS, to enable explicit control of ten attributes.
- Score: 31.64341800095214
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: State-of-the-art text simplification (TS) systems adopt end-to-end neural
network models to directly generate the simplified version of the input text,
and usually function as a blackbox. Moreover, TS is usually treated as an
all-purpose generic task under the assumption of homogeneity, where the same
simplification is suitable for all. In recent years, however, there has been
increasing recognition of the need to adapt the simplification techniques to
the specific needs of different target groups. In this work, we aim to advance
current research on explainable and controllable TS in two ways: First,
building on recently proposed work to increase the transparency of TS systems,
we use a large set of (psycho-)linguistic features in combination with
pre-trained language models to improve explainable complexity prediction.
Second, based on the results of this preliminary task, we extend a
state-of-the-art Seq2Seq TS model, ACCESS, to enable explicit control of ten
attributes. The results of experiments show (1) that our approach improves the
performance of state-of-the-art models for predicting explainable complexity
and (2) that explicitly conditioning the Seq2Seq model on ten attributes leads
to a significant improvement in performance in both within-domain and
out-of-domain settings.
Related papers
- Automatically Learning Hybrid Digital Twins of Dynamical Systems [56.69628749813084]
Digital Twins (DTs) simulate the states and temporal dynamics of real-world systems.
DTs often struggle to generalize to unseen conditions in data-scarce settings.
In this paper, we propose an evolutionary algorithm ($textbfHDTwinGen$) to autonomously propose, evaluate, and optimize HDTwins.
arXiv Detail & Related papers (2024-10-31T07:28:22Z) - LLM-TS Integrator: Integrating LLM for Enhanced Time Series Modeling [5.853711797849859]
Time series(TS) modeling is essential in dynamic systems like weather prediction and anomaly detection.
Recent studies utilize Large Language Models (LLMs) for TS modeling, leveraging their powerful pattern recognition capabilities.
arXiv Detail & Related papers (2024-10-21T20:29:46Z) - Layerwise complexity-matched learning yields an improved model of cortical area V2 [12.861402235256207]
Deep neural networks trained end-to-end for object recognition approach human capabilities.
We develop a self-supervised training methodology that operates independently on successive layers.
We show that our model is better aligned with selectivity properties and neural activity in primate area V2.
arXiv Detail & Related papers (2023-12-18T18:37:02Z) - When Parameter-efficient Tuning Meets General-purpose Vision-language
Models [65.19127815275307]
PETAL revolutionizes the training process by requiring only 0.5% of the total parameters, achieved through a unique mode approximation technique.
Our experiments reveal that PETAL not only outperforms current state-of-the-art methods in most scenarios but also surpasses full fine-tuning models in effectiveness.
arXiv Detail & Related papers (2023-12-16T17:13:08Z) - On Conditional and Compositional Language Model Differentiable Prompting [75.76546041094436]
Prompts have been shown to be an effective method to adapt a frozen Pretrained Language Model (PLM) to perform well on downstream tasks.
We propose a new model, Prompt Production System (PRopS), which learns to transform task instructions or input metadata, into continuous prompts.
arXiv Detail & Related papers (2023-07-04T02:47:42Z) - Enhancing Pre-trained Models with Text Structure Knowledge for Question
Generation [2.526624977753083]
We model text structure as answer position and syntactic dependency, and propose answer localness modeling and syntactic mask attention to address these limitations.
Experiments on SQuAD dataset show that our proposed two modules improve performance over the strong pre-trained model ProphetNet.
arXiv Detail & Related papers (2022-09-09T08:33:47Z) - Speech Enhancement with Score-Based Generative Models in the Complex
STFT Domain [18.090665052145653]
We propose a novel training task for speech enhancement using a complex-valued deep neural network.
We derive this training task within the formalism of differential equations, thereby enabling the use of predictor-corrector samplers.
arXiv Detail & Related papers (2022-03-31T12:53:47Z) - Factorized Neural Transducer for Efficient Language Model Adaptation [51.81097243306204]
We propose a novel model, factorized neural Transducer, by factorizing the blank and vocabulary prediction.
It is expected that this factorization can transfer the improvement of the standalone language model to the Transducer for speech recognition.
We demonstrate that the proposed factorized neural Transducer yields 15% to 20% WER improvements when out-of-domain text data is used for language model adaptation.
arXiv Detail & Related papers (2021-09-27T15:04:00Z) - A Full Text-Dependent End to End Mispronunciation Detection and
Diagnosis with Easy Data Augmentation Techniques [28.59181595057581]
We present a novel text-dependent model which is difference with sed-mdd.
We propose three simple data augmentation methods, which effectively improve the ability of model to capture mispronounced phonemes.
arXiv Detail & Related papers (2021-04-17T03:11:41Z) - SDA: Improving Text Generation with Self Data Augmentation [88.24594090105899]
We propose to improve the standard maximum likelihood estimation (MLE) paradigm by incorporating a self-imitation-learning phase for automatic data augmentation.
Unlike most existing sentence-level augmentation strategies, our method is more general and could be easily adapted to any MLE-based training procedure.
arXiv Detail & Related papers (2021-01-02T01:15:57Z) - Controllable Text Simplification with Explicit Paraphrasing [88.02804405275785]
Text Simplification improves the readability of sentences through several rewriting transformations, such as lexical paraphrasing, deletion, and splitting.
Current simplification systems are predominantly sequence-to-sequence models that are trained end-to-end to perform all these operations simultaneously.
We propose a novel hybrid approach that leverages linguistically-motivated rules for splitting and deletion, and couples them with a neural paraphrasing model to produce varied rewriting styles.
arXiv Detail & Related papers (2020-10-21T13:44:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.