Studying the role of named entities for content preservation in text
style transfer
- URL: http://arxiv.org/abs/2206.09676v1
- Date: Mon, 20 Jun 2022 09:31:47 GMT
- Title: Studying the role of named entities for content preservation in text
style transfer
- Authors: Nikolay Babakov, David Dale, Varvara Logacheva, Irina Krotova,
Alexander Panchenko
- Abstract summary: We focus on the role of named entities in content preservation for formality text style transfer.
We collect a new dataset for the evaluation of content similarity measures in text style transfer.
We perform an error analysis of a pre-trained formality transfer model and introduce a simple technique to use information about named entities to enhance the performance of baseline content similarity measures used in text style transfer.
- Score: 65.40394342240558
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Text style transfer techniques are gaining popularity in Natural Language
Processing, finding various applications such as text detoxification,
sentiment, or formality transfer. However, the majority of the existing
approaches were tested on such domains as online communications on public
platforms, music, or entertainment yet none of them were applied to the domains
which are typical for task-oriented production systems, such as personal plans
arrangements (e.g. booking of flights or reserving a table in a restaurant). We
fill this gap by studying formality transfer in this domain.
We noted that the texts in this domain are full of named entities, which are
very important for keeping the original sense of the text. Indeed, if for
example, someone communicates the destination city of a flight it must not be
altered. Thus, we concentrate on the role of named entities in content
preservation for formality text style transfer.
We collect a new dataset for the evaluation of content similarity measures in
text style transfer. It is taken from a corpus of task-oriented dialogues and
contains many important entities related to realistic requests that make this
dataset particularly useful for testing style transfer models before using them
in production. Besides, we perform an error analysis of a pre-trained formality
transfer model and introduce a simple technique to use information about named
entities to enhance the performance of baseline content similarity measures
used in text style transfer.
Related papers
- ParaGuide: Guided Diffusion Paraphrasers for Plug-and-Play Textual Style
Transfer [57.6482608202409]
Textual style transfer is the task of transforming stylistic properties of text while preserving meaning.
We introduce a novel diffusion-based framework for general-purpose style transfer that can be flexibly adapted to arbitrary target styles.
We validate the method on the Enron Email Corpus, with both human and automatic evaluations, and find that it outperforms strong baselines on formality, sentiment, and even authorship style transfer.
arXiv Detail & Related papers (2023-08-29T17:36:02Z) - Don't lose the message while paraphrasing: A study on content preserving
style transfer [61.38460184163704]
Content preservation is critical for real-world applications of style transfer studies.
We compare various style transfer models on the example of the formality transfer domain.
We conduct a precise comparative study of several state-of-the-art techniques for style transfer.
arXiv Detail & Related papers (2023-08-17T15:41:08Z) - Generalized One-shot Domain Adaption of Generative Adversarial Networks [72.84435077616135]
The adaption of Generative Adversarial Network (GAN) aims to transfer a pre-trained GAN to a given domain with limited training data.
We consider that the adaptation from source domain to target domain can be decoupled into two parts: the transfer of global style like texture and color, and the emergence of new entities that do not belong to the source domain.
Our core objective is to constrain the gap between the internal distributions of the reference and syntheses by sliced Wasserstein distance.
arXiv Detail & Related papers (2022-09-08T09:24:44Z) - StoryTrans: Non-Parallel Story Author-Style Transfer with Discourse
Representations and Content Enhancing [73.81778485157234]
Long texts usually involve more complicated author linguistic preferences such as discourse structures than sentences.
We formulate the task of non-parallel story author-style transfer, which requires transferring an input story into a specified author style.
We use an additional training objective to disentangle stylistic features from the learned discourse representation to prevent the model from degenerating to an auto-encoder.
arXiv Detail & Related papers (2022-08-29T08:47:49Z) - DSGPT: Domain-Specific Generative Pre-Training of Transformers for Text
Generation in E-commerce Title and Review Summarization [14.414693156937782]
We propose a novel domain-specific generative pre-training (DS-GPT) method for text generation.
We apply it to the product titleand review summarization problems on E-commerce mobile display.
arXiv Detail & Related papers (2021-12-15T19:02:49Z) - Reinforced Rewards Framework for Text Style Transfer [11.49063129923684]
Style transfer deals with the algorithms to transfer the stylistic properties of a piece of text into that of another while ensuring that the core content is preserved.
Existing works evaluate the style transfer models based on content preservation and transfer strength.
We propose a reinforcement learning based framework that directly rewards the framework on these target metrics yielding a better transfer of the target style.
arXiv Detail & Related papers (2020-05-11T16:54:28Z) - Exploring the Limits of Transfer Learning with a Unified Text-to-Text
Transformer [64.22926988297685]
Transfer learning, where a model is first pre-trained on a data-rich task before being fine-tuned on a downstream task, has emerged as a powerful technique in natural language processing (NLP)
In this paper, we explore the landscape of introducing transfer learning techniques for NLP by a unified framework that converts all text-based language problems into a text-to-text format.
arXiv Detail & Related papers (2019-10-23T17:37:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.