Unsupervised Text Style Transfer with Deep Generative Models
- URL: http://arxiv.org/abs/2308.16584v1
- Date: Thu, 31 Aug 2023 09:29:35 GMT
- Title: Unsupervised Text Style Transfer with Deep Generative Models
- Authors: Zhongtao Jiang, Yuanzhe Zhang, Yiming Ju, and Kang Liu
- Abstract summary: We present a general framework for unsupervised text style transfer with deep generative models.
Our framework is able to unify previous embedding and prototype methods as two special forms.
It also provides a principled perspective to explain previously proposed techniques in the field such as aligned encoder and adversarial training.
- Score: 12.801169425020225
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We present a general framework for unsupervised text style transfer with deep
generative models. The framework models each sentence-label pair in the
non-parallel corpus as partially observed from a complete quadruplet which
additionally contains two latent codes representing the content and style,
respectively. These codes are learned by exploiting dependencies inside the
observed data. Then a sentence is transferred by manipulating them. Our
framework is able to unify previous embedding and prototype methods as two
special forms. It also provides a principled perspective to explain previously
proposed techniques in the field such as aligned encoder and adversarial
training. We further conduct experiments on three benchmarks. Both automatic
and human evaluation results show that our methods achieve better or
competitive results compared to several strong baselines.
Related papers
- Prefix-Tuning Based Unsupervised Text Style Transfer [29.86587278794342]
Unsupervised text style transfer aims at training a generative model that can alter the style of the input sentence while preserving its content.
In this paper, we employ powerful pre-trained large language models and present a new prefix-tuning-based method for unsupervised text style transfer.
arXiv Detail & Related papers (2023-10-23T06:13:08Z) - Repurposing Knowledge Graph Embeddings for Triple Representation via
Weak Supervision [77.34726150561087]
Current methods learn triple embeddings from scratch without utilizing entity and predicate embeddings from pre-trained models.
We develop a method for automatically sampling triples from a knowledge graph and estimating their pairwise similarities from pre-trained embedding models.
These pairwise similarity scores are then fed to a Siamese-like neural architecture to fine-tune triple representations.
arXiv Detail & Related papers (2022-08-22T14:07:08Z) - Improving Tree-Structured Decoder Training for Code Generation via
Mutual Learning [27.080718377956693]
Code generation aims to automatically generate a piece of code given an input natural language utterance.
We first throughly analyze the context modeling difference between neural code generation models with different decodings.
We propose to introduce a mutual learning framework to jointly train these models.
arXiv Detail & Related papers (2021-05-31T08:44:13Z) - Towards Open-World Text-Guided Face Image Generation and Manipulation [52.83401421019309]
We propose a unified framework for both face image generation and manipulation.
Our method supports open-world scenarios, including both image and text, without any re-training, fine-tuning, or post-processing.
arXiv Detail & Related papers (2021-04-18T16:56:07Z) - Inference Time Style Control for Summarization [6.017006996402699]
We present two novel methods that can be deployed during summary decoding on any pre-trained Transformer-based summarization model.
In experiments of summarizing with simplicity control, automatic evaluation and human judges both find our models producing outputs in simpler languages while still informative.
arXiv Detail & Related papers (2021-04-05T00:27:18Z) - Hierarchical Bi-Directional Self-Attention Networks for Paper Review
Rating Recommendation [81.55533657694016]
We propose a Hierarchical bi-directional self-attention Network framework (HabNet) for paper review rating prediction and recommendation.
Specifically, we leverage the hierarchical structure of the paper reviews with three levels of encoders: sentence encoder (level one), intra-review encoder (level two) and inter-review encoder (level three)
We are able to identify useful predictors to make the final acceptance decision, as well as to help discover the inconsistency between numerical review ratings and text sentiment conveyed by reviewers.
arXiv Detail & Related papers (2020-11-02T08:07:50Z) - Unsupervised Paraphrasing with Pretrained Language Models [85.03373221588707]
We propose a training pipeline that enables pre-trained language models to generate high-quality paraphrases in an unsupervised setting.
Our recipe consists of task-adaptation, self-supervision, and a novel decoding algorithm named Dynamic Blocking.
We show with automatic and human evaluations that our approach achieves state-of-the-art performance on both the Quora Question Pair and the ParaNMT datasets.
arXiv Detail & Related papers (2020-10-24T11:55:28Z) - Narrative Text Generation with a Latent Discrete Plan [39.71663365273463]
We propose a deep latent variable model that first samples a sequence of anchor words, one per sentence in the story, as part of its generative process.
During training, our model treats the sequence of anchor words as a latent variable and attempts to induce anchoring sequences that help guide generation in an unsupervised fashion.
We conduct human evaluations which demonstrate that the stories produced by our model are rated better in comparison with baselines which do not consider story plans.
arXiv Detail & Related papers (2020-10-07T08:45:37Z) - Structure-Augmented Text Representation Learning for Efficient Knowledge
Graph Completion [53.31911669146451]
Human-curated knowledge graphs provide critical supportive information to various natural language processing tasks.
These graphs are usually incomplete, urging auto-completion of them.
graph embedding approaches, e.g., TransE, learn structured knowledge via representing graph elements into dense embeddings.
textual encoding approaches, e.g., KG-BERT, resort to graph triple's text and triple-level contextualized representations.
arXiv Detail & Related papers (2020-04-30T13:50:34Z) - A Probabilistic Formulation of Unsupervised Text Style Transfer [128.80213211598752]
We present a deep generative model for unsupervised text style transfer that unifies previously proposed non-generative techniques.
By hypothesizing a parallel latent sequence that generates each observed sequence, our model learns to transform sequences from one domain to another in a completely unsupervised fashion.
arXiv Detail & Related papers (2020-02-10T16:20:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.