A Quality-based Syntactic Template Retriever for
Syntactically-controlled Paraphrase Generation
- URL: http://arxiv.org/abs/2310.13262v1
- Date: Fri, 20 Oct 2023 03:55:39 GMT
- Title: A Quality-based Syntactic Template Retriever for
Syntactically-controlled Paraphrase Generation
- Authors: Xue Zhang, Songming Zhang, Yunlong Liang, Yufeng Chen, Jian Liu,
Wenjuan Han, Jinan Xu
- Abstract summary: Existing syntactically-controlled paraphrase generation models perform promisingly with human-annotated or well-chosen syntactic templates.
The prohibitive cost makes it unfeasible to manually design decent templates for every source sentence.
We propose a novel Quality-based Syntactic Template Retriever (QSTR) to retrieve templates based on the quality of the to-be-generated paraphrases.
- Score: 67.98367574025797
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing syntactically-controlled paraphrase generation (SPG) models perform
promisingly with human-annotated or well-chosen syntactic templates. However,
the difficulty of obtaining such templates actually hinders the practical
application of SPG models. For one thing, the prohibitive cost makes it
unfeasible to manually design decent templates for every source sentence. For
another, the templates automatically retrieved by current heuristic methods are
usually unreliable for SPG models to generate qualified paraphrases. To escape
this dilemma, we propose a novel Quality-based Syntactic Template Retriever
(QSTR) to retrieve templates based on the quality of the to-be-generated
paraphrases. Furthermore, for situations requiring multiple paraphrases for
each source sentence, we design a Diverse Templates Search (DTS) algorithm,
which can enhance the diversity between paraphrases without sacrificing
quality. Experiments demonstrate that QSTR can significantly surpass existing
retrieval methods in generating high-quality paraphrases and even perform
comparably with human-annotated templates in terms of reference-free metrics.
Additionally, human evaluation and the performance on downstream tasks using
our generated paraphrases for data augmentation showcase the potential of our
QSTR and DTS algorithm in practical scenarios.
Related papers
- Detection and Measurement of Syntactic Templates in Generated Text [58.111650675717414]
We offer an analysis of syntactic features to characterize general repetition in models.
We find that models tend to produce templated text in downstream tasks at a higher rate than what is found in human-reference texts.
arXiv Detail & Related papers (2024-06-28T19:34:23Z) - Retrieval is Accurate Generation [99.24267226311157]
We introduce a novel method that selects context-aware phrases from a collection of supporting documents.
Our model achieves the best performance and the lowest latency among several retrieval-augmented baselines.
arXiv Detail & Related papers (2024-02-27T14:16:19Z) - ParaAMR: A Large-Scale Syntactically Diverse Paraphrase Dataset by AMR
Back-Translation [59.91139600152296]
ParaAMR is a large-scale syntactically diverse paraphrase dataset created by abstract meaning representation back-translation.
We show that ParaAMR can be used to improve on three NLP tasks: learning sentence embeddings, syntactically controlled paraphrase generation, and data augmentation for few-shot learning.
arXiv Detail & Related papers (2023-05-26T02:27:33Z) - Unsupervised Syntactically Controlled Paraphrase Generation with
Abstract Meaning Representations [59.10748929158525]
Abstract Representations (AMR) can greatly improve the performance of unsupervised syntactically controlled paraphrase generation.
Our proposed model, AMR-enhanced Paraphrase Generator (AMRPG), encodes the AMR graph and the constituency parses the input sentence into two disentangled semantic and syntactic embeddings.
Experiments show that AMRPG generates more accurate syntactically controlled paraphrases, both quantitatively and qualitatively, compared to the existing unsupervised approaches.
arXiv Detail & Related papers (2022-11-02T04:58:38Z) - Template-free Prompt Tuning for Few-shot NER [46.59447116255979]
We propose a more elegant method to reformulate NER tasks as LM problems without any templates.
Specifically, we discard the template construction process while maintaining the word prediction paradigm of pre-training models.
Experimental results demonstrate the effectiveness of the proposed method over bert-tagger and template-based method under few-shot setting.
arXiv Detail & Related papers (2021-09-28T07:19:24Z) - Generating Syntactically Controlled Paraphrases without Using Annotated
Parallel Pairs [37.808235216195484]
We show that it is possible to generate syntactically various paraphrases without the need for annotated paraphrase pairs.
We propose Syntactically controlled Paraphrase Generator (SynPG), an encoder-decoder based model that learns to disentangle the semantics and the syntax of a sentence.
arXiv Detail & Related papers (2021-01-26T06:13:52Z) - Template Controllable keywords-to-text Generation [16.255080737147384]
The model takes as input a set of un-ordered keywords, and part-of-speech (POS) based template instructions.
The framework is based on the encode-attend-decode paradigm, where keywords and templates are encoded first, and the decoder judiciously attends over the contexts derived from the encoded keywords and templates to generate the sentences.
arXiv Detail & Related papers (2020-11-07T08:05:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.