Learning to Selectively Learn for Weakly-supervised Paraphrase
Generation
- URL: http://arxiv.org/abs/2109.12457v1
- Date: Sat, 25 Sep 2021 23:31:13 GMT
- Title: Learning to Selectively Learn for Weakly-supervised Paraphrase
Generation
- Authors: Kaize Ding, Dingcheng Li, Alexander Hanbo Li, Xing Fan, Chenlei Guo,
Yang Liu and Huan Liu
- Abstract summary: We propose a novel approach to generate high-quality paraphrases with weak supervision data.
Specifically, we tackle the weakly-supervised paraphrase generation problem by:.
obtaining abundant weakly-labeled parallel sentences via retrieval-based pseudo paraphrase expansion.
We demonstrate that our approach achieves significant improvements over existing unsupervised approaches, and is even comparable in performance with supervised state-of-the-arts.
- Score: 81.65399115750054
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Paraphrase generation is a longstanding NLP task that has diverse
applications for downstream NLP tasks. However, the effectiveness of existing
efforts predominantly relies on large amounts of golden labeled data. Though
unsupervised endeavors have been proposed to address this issue, they may fail
to generate meaningful paraphrases due to the lack of supervision signals. In
this work, we go beyond the existing paradigms and propose a novel approach to
generate high-quality paraphrases with weak supervision data. Specifically, we
tackle the weakly-supervised paraphrase generation problem by: (1) obtaining
abundant weakly-labeled parallel sentences via retrieval-based pseudo
paraphrase expansion; and (2) developing a meta-learning framework to
progressively select valuable samples for fine-tuning a pre-trained language
model, i.e., BART, on the sentential paraphrasing task. We demonstrate that our
approach achieves significant improvements over existing unsupervised
approaches, and is even comparable in performance with supervised
state-of-the-arts.
Related papers
- MarvelOVD: Marrying Object Recognition and Vision-Language Models for Robust Open-Vocabulary Object Detection [107.15164718585666]
We investigate the root cause of VLMs' biased prediction under the open vocabulary detection context.
Our observations lead to a simple yet effective paradigm, coded MarvelOVD, that generates significantly better training targets.
Our method outperforms the other state-of-the-arts by significant margins.
arXiv Detail & Related papers (2024-07-31T09:23:57Z) - Unsupervised Syntactically Controlled Paraphrase Generation with
Abstract Meaning Representations [59.10748929158525]
Abstract Representations (AMR) can greatly improve the performance of unsupervised syntactically controlled paraphrase generation.
Our proposed model, AMR-enhanced Paraphrase Generator (AMRPG), encodes the AMR graph and the constituency parses the input sentence into two disentangled semantic and syntactic embeddings.
Experiments show that AMRPG generates more accurate syntactically controlled paraphrases, both quantitatively and qualitatively, compared to the existing unsupervised approaches.
arXiv Detail & Related papers (2022-11-02T04:58:38Z) - Learning from Bootstrapping and Stepwise Reinforcement Reward: A
Semi-Supervised Framework for Text Style Transfer [30.622772801446132]
We propose a semi-supervised framework for text style transfer.
First, the learning process is bootstrapped with supervision guided by automatically constructed pseudo-parallel pairs.
Then the model learns from unlabeled data via reinforcement rewards.
arXiv Detail & Related papers (2022-05-19T05:18:06Z) - Entailment Relation Aware Paraphrase Generation [17.6146622291895]
We propose a reinforcement learning-based weakly-supervised paraphrasing system, ERAP, that can be trained using existing paraphrase and natural language inference corpora.
A combination of automated and human evaluations show that ERAP generates paraphrases conforming to the specified entailment relation.
arXiv Detail & Related papers (2022-03-20T08:02:09Z) - Adversarial GLUE: A Multi-Task Benchmark for Robustness Evaluation of
Language Models [86.02610674750345]
Adversarial GLUE (AdvGLUE) is a new multi-task benchmark to explore and evaluate the vulnerabilities of modern large-scale language models under various types of adversarial attacks.
We apply 14 adversarial attack methods to GLUE tasks to construct AdvGLUE, which is further validated by humans for reliable annotations.
All the language models and robust training methods we tested perform poorly on AdvGLUE, with scores lagging far behind the benign accuracy.
arXiv Detail & Related papers (2021-11-04T12:59:55Z) - ConRPG: Paraphrase Generation using Contexts as Regularizer [31.967883219986362]
A long-standing issue with paraphrase generation is how to obtain reliable supervision signals.
We propose an unsupervised paradigm for paraphrase generation based on the assumption that the probabilities of generating two sentences with the same meaning given the same context should be the same.
We propose a pipelined system which consists of paraphrase candidate generation based on contextual language models, candidate filtering using scoring functions, and paraphrase model training based on the selected candidates.
arXiv Detail & Related papers (2021-09-01T12:57:30Z) - Unsupervised Paraphrasing with Pretrained Language Models [85.03373221588707]
We propose a training pipeline that enables pre-trained language models to generate high-quality paraphrases in an unsupervised setting.
Our recipe consists of task-adaptation, self-supervision, and a novel decoding algorithm named Dynamic Blocking.
We show with automatic and human evaluations that our approach achieves state-of-the-art performance on both the Quora Question Pair and the ParaNMT datasets.
arXiv Detail & Related papers (2020-10-24T11:55:28Z) - Reflective Decoding: Beyond Unidirectional Generation with Off-the-Shelf
Language Models [63.808843089941405]
Large pretrained LanguageModels (LMs) generate text with remarkable quality, but only sequentially from left to right.
We present Reflective Decoding, a novel unsupervised algorithm that allows for direct application of unidirectional LMs to non-sequential tasks.
Our 2-step approach requires no supervision or even parallel corpora, only two off-the-shelf pretrained LMs in opposite directions.
arXiv Detail & Related papers (2020-10-16T18:02:07Z) - Unsupervised Paraphrase Generation using Pre-trained Language Models [0.0]
OpenAI's GPT-2 is notable for its capability to generate fluent, well formulated, grammatically consistent text.
We leverage this generation capability of GPT-2 to generate paraphrases without any supervision from labelled data.
Our experiments show that paraphrases generated with our model are of good quality, are diverse and improves the downstream task performance when used for data augmentation.
arXiv Detail & Related papers (2020-06-09T19:40:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.