Do Language Models Plagiarize?
- URL: http://arxiv.org/abs/2203.07618v1
- Date: Tue, 15 Mar 2022 03:11:11 GMT
- Title: Do Language Models Plagiarize?
- Authors: Jooyoung Lee, Thai Le, Jinghui Chen, Dongwon Lee
- Abstract summary: We investigate whether language models memorize but also plagiarize training samples when generating artificial texts.
Our findings support that they, especially GPT-2, reuse particular pieces of texts from the training corpus with or without obfuscation.
Our work implies that future research on neural language models should take precautions to avoid models plagiarizing their training datasets.
- Score: 22.02731537718498
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Past literature has illustrated that language models do not fully understand
the context and sensitivity of text and can sometimes memorize phrases or
sentences present in their training sets. In this paper, we investigate whether
they not only memorize but also plagiarize training samples when generating
artificial texts. Our findings support that they, especially GPT-2, reuse
particular pieces of texts from the training corpus with or without
obfuscation. We have four main results: 1) language models with more capacity
plagiarize more; 2) fine-tuned language models demonstrate differing patterns
of plagiarism based on characteristics of auxiliary data; 3) sampling from
truncated language modeling distributions tends to heighten the degree of
plagiarism as opposed to temperature sampling, and 4) plagiarism in language
models can have serious privacy consequences. Overall, our work implies that
future research on neural language models should take precautions to avoid
models plagiarizing their training datasets.
Related papers
- Measuring Non-Adversarial Reproduction of Training Data in Large Language Models [71.55350441396243]
We quantify the overlap between model responses and pretraining data when responding to natural and benign prompts.
We find that up to 15% of the text output by popular conversational language models overlaps with snippets from the Internet.
While appropriate prompting can reduce non-adversarial reproduction on average, we find that mitigating worst-case reproduction of training data requires stronger defenses.
arXiv Detail & Related papers (2024-11-15T14:55:01Z) - FOCUS: Forging Originality through Contrastive Use in Self-Plagiarism for Language Models [38.76912842622624]
Pre-trained Language Models (PLMs) have shown impressive results in various Natural Language Generation (NLG) tasks.
This study introduces a unique "self-plagiarism" contrastive decoding strategy, aimed at boosting the originality of text produced by PLMs.
arXiv Detail & Related papers (2024-06-02T19:17:00Z) - Detecting Mode Collapse in Language Models via Narration [0.0]
We study 4,374 stories sampled from three OpenAI language models.
We show successive versions of GPT-3 suffer from increasing degrees of "mode collapse"
Our method and results are significant for researchers seeking to employ language models in sociological simulations.
arXiv Detail & Related papers (2024-02-06T23:52:58Z) - Few-Shot Detection of Machine-Generated Text using Style Representations [4.326503887981912]
Language models that convincingly mimic human writing pose a significant risk of abuse.
We propose to leverage representations of writing style estimated from human-authored text.
We find that features effective at distinguishing among human authors are also effective at distinguishing human from machine authors.
arXiv Detail & Related papers (2024-01-12T17:26:51Z) - Fine-tuning Language Models for Factuality [96.5203774943198]
Large pre-trained language models (LLMs) have led to their widespread use, sometimes even as a replacement for traditional search engines.
Yet language models are prone to making convincing but factually inaccurate claims, often referred to as 'hallucinations'
In this work, we fine-tune language models to be more factual, without human labeling.
arXiv Detail & Related papers (2023-11-14T18:59:15Z) - Forgetting Private Textual Sequences in Language Models via
Leave-One-Out Ensemble [13.893379594151533]
We propose a novel leave-one-out ensemble method to unlearn the targeted textual sequences that need to be forgotten from the model.
Experiments on LibriSpeech and WikiText-103 datasets show that the proposed method achieves superior privacy-utility trade-offs than other counterparts.
arXiv Detail & Related papers (2023-09-28T00:43:18Z) - Preventing Verbatim Memorization in Language Models Gives a False Sense
of Privacy [91.98116450958331]
We argue that verbatim memorization definitions are too restrictive and fail to capture more subtle forms of memorization.
Specifically, we design and implement an efficient defense that perfectly prevents all verbatim memorization.
We conclude by discussing potential alternative definitions and why defining memorization is a difficult yet crucial open question for neural language models.
arXiv Detail & Related papers (2022-10-31T17:57:55Z) - How much do language models copy from their training data? Evaluating
linguistic novelty in text generation using RAVEN [63.79300884115027]
Current language models can generate high-quality text.
Are they simply copying text they have seen before, or have they learned generalizable linguistic abstractions?
We introduce RAVEN, a suite of analyses for assessing the novelty of generated text.
arXiv Detail & Related papers (2021-11-18T04:07:09Z) - Extracting Training Data from Large Language Models [78.3839333127544]
This paper demonstrates that an adversary can perform a training data extraction attack to recover individual training examples by querying the language model.
We demonstrate our attack on GPT-2, a language model trained on scrapes of the public Internet, and are able to extract hundreds of verbatim text sequences from the model's training data.
arXiv Detail & Related papers (2020-12-14T18:39:09Z) - Exemplar-Controllable Paraphrasing and Translation using Bitext [57.92051459102902]
We adapt models from prior work to be able to learn solely from bilingual text (bitext)
Our single proposed model can perform four tasks: controlled paraphrase generation in both languages and controlled machine translation in both language directions.
arXiv Detail & Related papers (2020-10-12T17:02:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.