Replacing Language Model for Style Transfer
- URL: http://arxiv.org/abs/2211.07343v2
- Date: Wed, 28 Feb 2024 12:51:09 GMT
- Title: Replacing Language Model for Style Transfer
- Authors: Pengyu Cheng, Ruineng Li
- Abstract summary: We introduce replacing language model (RLM), a sequence-to-sequence language modeling framework for text style transfer (TST)
Our method autoregressively replaces each token of the source sentence with a text span that has a similar meaning but in the target style.
The new span is generated via a non-autoregressive masked language model, which can better preserve the local-contextual meaning of the replaced token.
- Score: 6.364517234783756
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce replacing language model (RLM), a sequence-to-sequence language
modeling framework for text style transfer (TST). Our method autoregressively
replaces each token of the source sentence with a text span that has a similar
meaning but in the target style. The new span is generated via a
non-autoregressive masked language model, which can better preserve the
local-contextual meaning of the replaced token. This RLM generation scheme
gathers the flexibility of autoregressive models and the accuracy of
non-autoregressive models, which bridges the gap between sentence-level and
word-level style transfer methods. To control the generation style more
precisely, we conduct a token-level style-content disentanglement on the hidden
representations of RLM. Empirical results on real-world text datasets
demonstrate the effectiveness of RLM compared with other TST baselines. The
code is at https://github.com/Linear95/RLM.
Related papers
- TransMI: A Framework to Create Strong Baselines from Multilingual Pretrained Language Models for Transliterated Data [50.40191599304911]
We propose Transliterate transliteration-Merge (TransMI), which can create a strong baseline well-suited for data that is transliterated into a common script.
Results show a consistent improvement of 3% to 34%, varying across different models and tasks.
arXiv Detail & Related papers (2024-05-16T09:08:09Z) - Unsupervised Text Style Transfer via LLMs and Attention Masking with
Multi-way Interactions [18.64326057581588]
Unsupervised Text Style Transfer (UTST) has emerged as a critical task within the domain of Natural Language Processing (NLP)
We propose four ways of interactions, that are pipeline framework with tuned orders; knowledge distillation from Large Language Models (LLMs) to attention masking model; in-context learning with constructed parallel examples.
We empirically show these multi-way interactions can improve the baselines in certain perspective of style strength, content preservation and text fluency.
arXiv Detail & Related papers (2024-02-21T09:28:02Z) - FLIP: Fine-grained Alignment between ID-based Models and Pretrained Language Models for CTR Prediction [49.510163437116645]
Click-through rate (CTR) prediction plays as a core function module in personalized online services.
Traditional ID-based models for CTR prediction take as inputs the one-hot encoded ID features of tabular modality.
Pretrained Language Models(PLMs) has given rise to another paradigm, which takes as inputs the sentences of textual modality.
We propose to conduct Fine-grained feature-level ALignment between ID-based Models and Pretrained Language Models(FLIP) for CTR prediction.
arXiv Detail & Related papers (2023-10-30T11:25:03Z) - On Text Style Transfer via Style Masked Language Models [5.754152248672319]
Text Style Transfer (TST) is performable through approaches such as latent space disentanglement, cycleconsistency losses, prototype editing.
We present a prototype editing approach, which involves two key phases a) Masking of source style-associated tokens and b) Reconstruction of this source-style masked sentence conditioned with the target style.
We empirically show that this non-generational approach well suites the "content preserving" criteria for a task like TST, even for a complex baseline like Discourse.
arXiv Detail & Related papers (2022-10-12T16:44:06Z) - Collocation2Text: Controllable Text Generation from Guide Phrases in
Russian [0.0]
Collocation2Text is a plug-and-play method for automatic controllable text generation in Russian.
The method is based on two interacting models: the autoregressive language ruGPT-3 model and the autoencoding language ruRoBERTa model.
Experiments on generating news articles using the proposed method showed its effectiveness for automatically generated fluent texts.
arXiv Detail & Related papers (2022-06-18T17:10:08Z) - Text Detoxification using Large Pre-trained Neural Models [57.72086777177844]
We present two novel unsupervised methods for eliminating toxicity in text.
First method combines guidance of the generation process with small style-conditional language models.
Second method uses BERT to replace toxic words with their non-offensive synonyms.
arXiv Detail & Related papers (2021-09-18T11:55:32Z) - LT-LM: a novel non-autoregressive language model for single-shot lattice
rescoring [55.16665077221941]
We propose a novel rescoring approach, which processes the entire lattice in a single call to the model.
The key feature of our rescoring policy is a novel non-autoregressive Lattice Transformer Language Model (LT-LM)
arXiv Detail & Related papers (2021-04-06T14:06:07Z) - Unsupervised Text Style Transfer with Padded Masked Language Models [25.397832729384064]
Masker is an unsupervised text-editing method for style transfer.
It performs competitively in a fully unsupervised setting.
It improves supervised methods' accuracy by over 10 percentage points in low-resource settings.
arXiv Detail & Related papers (2020-10-02T15:33:42Z) - POINTER: Constrained Progressive Text Generation via Insertion-based
Generative Pre-training [93.79766670391618]
We present POINTER, a novel insertion-based approach for hard-constrained text generation.
The proposed method operates by progressively inserting new tokens between existing tokens in a parallel manner.
The resulting coarse-to-fine hierarchy makes the generation process intuitive and interpretable.
arXiv Detail & Related papers (2020-05-01T18:11:54Z) - PALM: Pre-training an Autoencoding&Autoregressive Language Model for
Context-conditioned Generation [92.7366819044397]
Self-supervised pre-training has emerged as a powerful technique for natural language understanding and generation.
This work presents PALM with a novel scheme that jointly pre-trains an autoencoding and autoregressive language model on a large unlabeled corpus.
An extensive set of experiments show that PALM achieves new state-of-the-art results on a variety of language generation benchmarks.
arXiv Detail & Related papers (2020-04-14T06:25:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.