Syntax-guided Controlled Generation of Paraphrases
- URL: http://arxiv.org/abs/2005.08417v1
- Date: Mon, 18 May 2020 01:31:28 GMT
- Title: Syntax-guided Controlled Generation of Paraphrases
- Authors: Ashutosh Kumar, Kabir Ahuja, Raghuram Vadapalli, Partha Talukdar
- Abstract summary: We propose Syntax Guided Controlled Paraphraser (SGCP), an end-to-end framework for syntactic paraphrase generation.
We find that SGCP can generate syntax conforming sentences while not compromising on relevance.
To drive future research, we have made SGCP's source code available.
- Score: 3.4129083593356433
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Given a sentence (e.g., "I like mangoes") and a constraint (e.g., sentiment
flip), the goal of controlled text generation is to produce a sentence that
adapts the input sentence to meet the requirements of the constraint (e.g., "I
hate mangoes"). Going beyond such simple constraints, recent works have started
exploring the incorporation of complex syntactic-guidance as constraints in the
task of controlled paraphrase generation. In these methods, syntactic-guidance
is sourced from a separate exemplar sentence. However, these prior works have
only utilized limited syntactic information available in the parse tree of the
exemplar sentence. We address this limitation in the paper and propose Syntax
Guided Controlled Paraphraser (SGCP), an end-to-end framework for syntactic
paraphrase generation. We find that SGCP can generate syntax conforming
sentences while not compromising on relevance. We perform extensive automated
and human evaluations over multiple real-world English language datasets to
demonstrate the efficacy of SGCP over state-of-the-art baselines. To drive
future research, we have made SGCP's source code available
Related papers
- Integrating Supertag Features into Neural Discontinuous Constituent Parsing [0.0]
Traditional views of constituency demand that constituents consist of adjacent words, common in languages like German.
Transition-based parsing produces trees given raw text input using supervised learning on large annotated corpora.
arXiv Detail & Related papers (2024-10-11T12:28:26Z) - Explicating the Implicit: Argument Detection Beyond Sentence Boundaries [24.728886446551577]
We reformulate the problem of argument detection through textual entailment to capture semantic relations across sentence boundaries.
Our method does not require direct supervision, which is generally absent due to dataset scarcity.
We demonstrate it on a recent document-level benchmark, outperforming some supervised methods and contemporary language models.
arXiv Detail & Related papers (2024-08-08T06:18:24Z) - Intertwining CP and NLP: The Generation of Unreasonably Constrained Sentences [49.86129209397701]
This paper presents the Constraints First Framework to remedy this issue.
It is solved by a constraint programming method that combines linguistic properties with more classical constraints.
The effectiveness of this approach is demonstrated by tackling a new more tediously constrained text generation problem.
arXiv Detail & Related papers (2024-06-15T17:40:49Z) - Language Model Sentence Completion with a Parser-Driven Rhetorical
Control Method [8.430481660019451]
Controlled text generation (CTG) seeks to guide large language model (LLM) output to produce text that conforms to desired criteria.
The current study presents a novel CTG algorithm that enforces adherence toward specific rhetorical relations.
The method is validated both with automatic and human evaluation.
arXiv Detail & Related papers (2024-02-09T01:15:42Z) - ParaAMR: A Large-Scale Syntactically Diverse Paraphrase Dataset by AMR
Back-Translation [59.91139600152296]
ParaAMR is a large-scale syntactically diverse paraphrase dataset created by abstract meaning representation back-translation.
We show that ParaAMR can be used to improve on three NLP tasks: learning sentence embeddings, syntactically controlled paraphrase generation, and data augmentation for few-shot learning.
arXiv Detail & Related papers (2023-05-26T02:27:33Z) - Unsupervised Syntactically Controlled Paraphrase Generation with
Abstract Meaning Representations [59.10748929158525]
Abstract Representations (AMR) can greatly improve the performance of unsupervised syntactically controlled paraphrase generation.
Our proposed model, AMR-enhanced Paraphrase Generator (AMRPG), encodes the AMR graph and the constituency parses the input sentence into two disentangled semantic and syntactic embeddings.
Experiments show that AMRPG generates more accurate syntactically controlled paraphrases, both quantitatively and qualitatively, compared to the existing unsupervised approaches.
arXiv Detail & Related papers (2022-11-02T04:58:38Z) - Towards Document-Level Paraphrase Generation with Sentence Rewriting and
Reordering [88.08581016329398]
We propose CoRPG (Coherence Relationship guided Paraphrase Generation) for document-level paraphrase generation.
We use graph GRU to encode the coherence relationship graph and get the coherence-aware representation for each sentence.
Our model can generate document paraphrase with more diversity and semantic preservation.
arXiv Detail & Related papers (2021-09-15T05:53:40Z) - Generating Syntactically Controlled Paraphrases without Using Annotated
Parallel Pairs [37.808235216195484]
We show that it is possible to generate syntactically various paraphrases without the need for annotated paraphrase pairs.
We propose Syntactically controlled Paraphrase Generator (SynPG), an encoder-decoder based model that learns to disentangle the semantics and the syntax of a sentence.
arXiv Detail & Related papers (2021-01-26T06:13:52Z) - Exemplar-Controllable Paraphrasing and Translation using Bitext [57.92051459102902]
We adapt models from prior work to be able to learn solely from bilingual text (bitext)
Our single proposed model can perform four tasks: controlled paraphrase generation in both languages and controlled machine translation in both language directions.
arXiv Detail & Related papers (2020-10-12T17:02:50Z) - Neural Syntactic Preordering for Controlled Paraphrase Generation [57.5316011554622]
Our work uses syntactic transformations to softly "reorder'' the source sentence and guide our neural paraphrasing model.
First, given an input sentence, we derive a set of feasible syntactic rearrangements using an encoder-decoder model.
Next, we use each proposed rearrangement to produce a sequence of position embeddings, which encourages our final encoder-decoder paraphrase model to attend to the source words in a particular order.
arXiv Detail & Related papers (2020-05-05T09:02:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.