Don't Change Me! User-Controllable Selective Paraphrase Generation
- URL: http://arxiv.org/abs/2008.09290v2
- Date: Mon, 25 Jan 2021 20:09:21 GMT
- Title: Don't Change Me! User-Controllable Selective Paraphrase Generation
- Authors: Mohan Zhang, Luchen Tan, Zhengkai Tu, Zihang Fu, Kun Xiong, Ming Li,
Jimmy Lin
- Abstract summary: In paraphrase generation, source sentences often contain phrases that should not be altered.
Our solution is to provide the user with explicit tags that can be placed around any arbitrary segment of text to mean "don't change me!"
The contribution of this work is a novel data generation technique using distant supervision.
- Score: 45.0436584774495
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the paraphrase generation task, source sentences often contain phrases
that should not be altered. Which phrases, however, can be context dependent
and can vary by application. Our solution to this challenge is to provide the
user with explicit tags that can be placed around any arbitrary segment of text
to mean "don't change me!" when generating a paraphrase; the model learns to
explicitly copy these phrases to the output. The contribution of this work is a
novel data generation technique using distant supervision that allows us to
start with a pretrained sequence-to-sequence model and fine-tune a paraphrase
generator that exhibits this behavior, allowing user-controllable paraphrase
generation. Additionally, we modify the loss during fine-tuning to explicitly
encourage diversity in model output. Our technique is language agnostic, and we
report experiments in English and Chinese.
Related papers
- Are Paraphrases Generated by Large Language Models Invertible? [4.148732457277201]
We consider the problem of paraphrase inversion: given a paraphrased document, attempt to recover the original text.
We fine-tune paraphrase inversion models, both with and without additional author-specific context.
We show that, when starting from paraphrased machine-generated text, we can recover significant portions of the document using a learned inversion model.
arXiv Detail & Related papers (2024-10-29T00:46:24Z) - Paraphrase Types for Generation and Detection [7.800428507692341]
We name these tasks Paraphrase Type Generation and Paraphrase Type Detection.
Our results suggest that while current techniques perform well in a binary classification scenario, the inclusion of fine-grained paraphrase types poses a significant challenge.
We believe paraphrase types can unlock a new paradigm for developing paraphrase models and solving tasks in the future.
arXiv Detail & Related papers (2023-10-23T12:32:41Z) - A Plug-and-Play Method for Controlled Text Generation [38.283313068622085]
We present a plug-and-play decoding method for controlled language generation that is so simple and intuitive, it can be described in a single sentence.
Despite the simplicity of this approach, we see it works incredibly well in practice.
arXiv Detail & Related papers (2021-09-20T17:27:03Z) - Towards Document-Level Paraphrase Generation with Sentence Rewriting and
Reordering [88.08581016329398]
We propose CoRPG (Coherence Relationship guided Paraphrase Generation) for document-level paraphrase generation.
We use graph GRU to encode the coherence relationship graph and get the coherence-aware representation for each sentence.
Our model can generate document paraphrase with more diversity and semantic preservation.
arXiv Detail & Related papers (2021-09-15T05:53:40Z) - From Solving a Problem Boldly to Cutting the Gordian Knot: Idiomatic
Text Generation [14.360808219541752]
We study a new application for text generation -- idiomatic sentence generation.
We propose a novel approach for this task, which retrieves the appropriate idiom for a given literal sentence.
We generate the idiomatic sentence by using a neural model to combine the retrieved idiom and the remainder of the sentence.
arXiv Detail & Related papers (2021-04-13T22:57:25Z) - Generating Adversarial Examples in Chinese Texts Using Sentence-Pieces [60.58900627906269]
We propose a pre-train language model as the substitutes generator using sentence-pieces to craft adversarial examples in Chinese.
The substitutions in the generated adversarial examples are not characters or words but textit'pieces', which are more natural to Chinese readers.
arXiv Detail & Related papers (2020-12-29T14:28:07Z) - Exemplar-Controllable Paraphrasing and Translation using Bitext [57.92051459102902]
We adapt models from prior work to be able to learn solely from bilingual text (bitext)
Our single proposed model can perform four tasks: controlled paraphrase generation in both languages and controlled machine translation in both language directions.
arXiv Detail & Related papers (2020-10-12T17:02:50Z) - Neural Syntactic Preordering for Controlled Paraphrase Generation [57.5316011554622]
Our work uses syntactic transformations to softly "reorder'' the source sentence and guide our neural paraphrasing model.
First, given an input sentence, we derive a set of feasible syntactic rearrangements using an encoder-decoder model.
Next, we use each proposed rearrangement to produce a sequence of position embeddings, which encourages our final encoder-decoder paraphrase model to attend to the source words in a particular order.
arXiv Detail & Related papers (2020-05-05T09:02:25Z) - Metaphoric Paraphrase Generation [58.592750281138265]
We use crowdsourcing to evaluate our results, as well as developing an automatic metric for evaluating metaphoric paraphrases.
We show that while the lexical replacement baseline is capable of producing accurate paraphrases, they often lack metaphoricity.
Our metaphor masking model excels in generating metaphoric sentences while performing nearly as well with regard to fluency and paraphrase quality.
arXiv Detail & Related papers (2020-02-28T16:30:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.