Replacing Language Model for Style Transfer
- URL: http://arxiv.org/abs/2211.07343v2
- Date: Wed, 28 Feb 2024 12:51:09 GMT
- Title: Replacing Language Model for Style Transfer
- Authors: Pengyu Cheng, Ruineng Li
- Abstract summary: We introduce replacing language model (RLM), a sequence-to-sequence language modeling framework for text style transfer (TST)
Our method autoregressively replaces each token of the source sentence with a text span that has a similar meaning but in the target style.
The new span is generated via a non-autoregressive masked language model, which can better preserve the local-contextual meaning of the replaced token.
- Score: 6.364517234783756
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We introduce replacing language model (RLM), a sequence-to-sequence language
modeling framework for text style transfer (TST). Our method autoregressively
replaces each token of the source sentence with a text span that has a similar
meaning but in the target style. The new span is generated via a
non-autoregressive masked language model, which can better preserve the
local-contextual meaning of the replaced token. This RLM generation scheme
gathers the flexibility of autoregressive models and the accuracy of
non-autoregressive models, which bridges the gap between sentence-level and
word-level style transfer methods. To control the generation style more
precisely, we conduct a token-level style-content disentanglement on the hidden
representations of RLM. Empirical results on real-world text datasets
demonstrate the effectiveness of RLM compared with other TST baselines. The
code is at https://github.com/Linear95/RLM.
Related papers
- DRESSing Up LLM: Efficient Stylized Question-Answering via Style Subspace Editing [23.467409274256255]
DRESS is a novel approach for generating stylized large language model (LLM) responses through representation editing.
Our approach disentangles a style-relevant subspace within the model's representation space to conduct representation editing.
In short, DRESS is a lightweight, train-free solution for enhancing LLMs with flexible and effective style control.
arXiv Detail & Related papers (2025-01-24T10:04:53Z) - Unsupervised Text Style Transfer via LLMs and Attention Masking with
Multi-way Interactions [18.64326057581588]
Unsupervised Text Style Transfer (UTST) has emerged as a critical task within the domain of Natural Language Processing (NLP)
We propose four ways of interactions, that are pipeline framework with tuned orders; knowledge distillation from Large Language Models (LLMs) to attention masking model; in-context learning with constructed parallel examples.
We empirically show these multi-way interactions can improve the baselines in certain perspective of style strength, content preservation and text fluency.
arXiv Detail & Related papers (2024-02-21T09:28:02Z) - FLIP: Fine-grained Alignment between ID-based Models and Pretrained Language Models for CTR Prediction [49.510163437116645]
Click-through rate (CTR) prediction plays as a core function module in personalized online services.
Traditional ID-based models for CTR prediction take as inputs the one-hot encoded ID features of tabular modality.
Pretrained Language Models(PLMs) has given rise to another paradigm, which takes as inputs the sentences of textual modality.
We propose to conduct Fine-grained feature-level ALignment between ID-based Models and Pretrained Language Models(FLIP) for CTR prediction.
arXiv Detail & Related papers (2023-10-30T11:25:03Z) - ParaGuide: Guided Diffusion Paraphrasers for Plug-and-Play Textual Style
Transfer [57.6482608202409]
Textual style transfer is the task of transforming stylistic properties of text while preserving meaning.
We introduce a novel diffusion-based framework for general-purpose style transfer that can be flexibly adapted to arbitrary target styles.
We validate the method on the Enron Email Corpus, with both human and automatic evaluations, and find that it outperforms strong baselines on formality, sentiment, and even authorship style transfer.
arXiv Detail & Related papers (2023-08-29T17:36:02Z) - On Text Style Transfer via Style Masked Language Models [5.754152248672319]
Text Style Transfer (TST) is performable through approaches such as latent space disentanglement, cycleconsistency losses, prototype editing.
We present a prototype editing approach, which involves two key phases a) Masking of source style-associated tokens and b) Reconstruction of this source-style masked sentence conditioned with the target style.
We empirically show that this non-generational approach well suites the "content preserving" criteria for a task like TST, even for a complex baseline like Discourse.
arXiv Detail & Related papers (2022-10-12T16:44:06Z) - Text Detoxification using Large Pre-trained Neural Models [57.72086777177844]
We present two novel unsupervised methods for eliminating toxicity in text.
First method combines guidance of the generation process with small style-conditional language models.
Second method uses BERT to replace toxic words with their non-offensive synonyms.
arXiv Detail & Related papers (2021-09-18T11:55:32Z) - LT-LM: a novel non-autoregressive language model for single-shot lattice
rescoring [55.16665077221941]
We propose a novel rescoring approach, which processes the entire lattice in a single call to the model.
The key feature of our rescoring policy is a novel non-autoregressive Lattice Transformer Language Model (LT-LM)
arXiv Detail & Related papers (2021-04-06T14:06:07Z) - Unsupervised Text Style Transfer with Padded Masked Language Models [25.397832729384064]
Masker is an unsupervised text-editing method for style transfer.
It performs competitively in a fully unsupervised setting.
It improves supervised methods' accuracy by over 10 percentage points in low-resource settings.
arXiv Detail & Related papers (2020-10-02T15:33:42Z) - POINTER: Constrained Progressive Text Generation via Insertion-based
Generative Pre-training [93.79766670391618]
We present POINTER, a novel insertion-based approach for hard-constrained text generation.
The proposed method operates by progressively inserting new tokens between existing tokens in a parallel manner.
The resulting coarse-to-fine hierarchy makes the generation process intuitive and interpretable.
arXiv Detail & Related papers (2020-05-01T18:11:54Z) - PALM: Pre-training an Autoencoding&Autoregressive Language Model for
Context-conditioned Generation [92.7366819044397]
Self-supervised pre-training has emerged as a powerful technique for natural language understanding and generation.
This work presents PALM with a novel scheme that jointly pre-trains an autoencoding and autoregressive language model on a large unlabeled corpus.
An extensive set of experiments show that PALM achieves new state-of-the-art results on a variety of language generation benchmarks.
arXiv Detail & Related papers (2020-04-14T06:25:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.