PEER: A Collaborative Language Model
- URL: http://arxiv.org/abs/2208.11663v1
- Date: Wed, 24 Aug 2022 16:56:47 GMT
- Title: PEER: A Collaborative Language Model
- Authors: Timo Schick, Jane Dwivedi-Yu, Zhengbao Jiang, Fabio Petroni, Patrick
Lewis, Gautier Izacard, Qingfei You, Christoforos Nalmpantis, Edouard Grave,
Sebastian Riedel
- Abstract summary: We introduce PEER, a collaborative language model that imitates the entire writing process itself.
PEER can write drafts, add suggestions, propose edits and provide explanations for its actions.
We show that PEER achieves strong performance across various domains and editing tasks.
- Score: 70.11876901409906
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Textual content is often the output of a collaborative writing process: We
start with an initial draft, ask for suggestions, and repeatedly make changes.
Agnostic of this process, today's language models are trained to generate only
the final result. As a consequence, they lack several abilities crucial for
collaborative writing: They are unable to update existing texts, difficult to
control and incapable of verbally planning or explaining their actions. To
address these shortcomings, we introduce PEER, a collaborative language model
that is trained to imitate the entire writing process itself: PEER can write
drafts, add suggestions, propose edits and provide explanations for its
actions. Crucially, we train multiple instances of PEER able to infill various
parts of the writing process, enabling the use of self-training techniques for
increasing the quality, amount and diversity of training data. This unlocks
PEER's full potential by making it applicable in domains for which no edit
histories are available and improving its ability to follow instructions, to
write useful comments, and to explain its actions. We show that PEER achieves
strong performance across various domains and editing tasks.
Related papers
- GhostWriter: Augmenting Collaborative Human-AI Writing Experiences
Through Personalization and Agency [1.7707677585873678]
Large language models (LLMs) are becoming more prevalent and have found a ubiquitous use in providing different forms of writing assistance.
We introduce GhostWriter, an AI-enhanced writing design probe where users can exercise enhanced agency and personalization.
We study 18 participants who use GhostWriter on two different writing tasks, observing that it helps users craft personalized text generations and empowers them by providing multiple ways to control the system's writing style.
arXiv Detail & Related papers (2024-02-13T23:48:59Z) - On the Robustness of Editing Large Language Models [57.477943944826904]
Large language models (LLMs) have played a pivotal role in building communicative AI, yet they encounter the challenge of efficient updates.
This work seeks to understand the strengths and limitations of editing methods, facilitating practical applications of communicative AI.
arXiv Detail & Related papers (2024-02-08T17:06:45Z) - ABScribe: Rapid Exploration & Organization of Multiple Writing Variations in Human-AI Co-Writing Tasks using Large Language Models [24.825435085579937]
We present ABScribe, an interface that supports rapid, yet visually structured, exploration and organization of writing variations.
With ABScribe, users can swiftly modify variations using LLM prompts, which are auto-converted into reusable buttons.
Variations are stored adjacently within text fields for rapid in-place comparisons using mouse-over interactions on a popup toolbar.
arXiv Detail & Related papers (2023-09-29T20:11:15Z) - PerPLM: Personalized Fine-tuning of Pretrained Language Models via
Writer-specific Intermediate Learning and Prompts [16.59511985633798]
Pretrained language models (PLMs) are powerful tools for capturing context.
PLMs are typically pretrained and fine-tuned for universal use across different writers.
This study aims to improve the accuracy of text understanding tasks by personalizing the fine-tuning of PLMs for specific writers.
arXiv Detail & Related papers (2023-09-14T14:03:48Z) - Stabilized In-Context Learning with Pre-trained Language Models for Few
Shot Dialogue State Tracking [57.92608483099916]
Large pre-trained language models (PLMs) have shown impressive unaided performance across many NLP tasks.
For more complex tasks such as dialogue state tracking (DST), designing prompts that reliably convey the desired intent is nontrivial.
We introduce a saliency model to limit dialogue text length, allowing us to include more exemplars per query.
arXiv Detail & Related papers (2023-02-12T15:05:10Z) - EditEval: An Instruction-Based Benchmark for Text Improvements [73.5918084416016]
This work presents EditEval: An instruction-based, benchmark and evaluation suite for automatic evaluation of editing capabilities.
We evaluate several pre-trained models, which shows that InstructGPT and PEER perform the best, but that most baselines fall below the supervised SOTA.
Our analysis shows that commonly used metrics for editing tasks do not always correlate well, and that optimization for prompts with the highest performance does not necessarily entail the strongest robustness to different models.
arXiv Detail & Related papers (2022-09-27T12:26:05Z) - CoditT5: Pretraining for Source Code and Natural Language Editing [34.77621217370665]
CoditT5 is a large language model for software-related editing tasks that is pretrained on large amounts of source code and natural language comments.
We fine-tune it on various downstream editing tasks, including comment updating, bug fixing, and automated code review.
arXiv Detail & Related papers (2022-08-10T16:59:40Z) - Read, Revise, Repeat: A System Demonstration for Human-in-the-loop
Iterative Text Revision [11.495407637511878]
We present a human-in-the-loop iterative text revision system, Read, Revise, Repeat (R3)
R3 aims at achieving high quality text revisions with minimal human efforts by reading model-generated revisions and user feedbacks, revising documents, and repeating human-machine interactions.
arXiv Detail & Related papers (2022-04-07T18:33:10Z) - CINS: Comprehensive Instruction for Few-shot Learning in Task-oriented
Dialog Systems [56.302581679816775]
This paper proposes Comprehensive Instruction (CINS) that exploits PLMs with task-specific instructions.
We design a schema (definition, constraint, prompt) of instructions and their customized realizations for three important downstream tasks in ToD.
Experiments are conducted on these ToD tasks in realistic few-shot learning scenarios with small validation data.
arXiv Detail & Related papers (2021-09-10T03:23:06Z) - Pre-training via Paraphrasing [96.79972492585112]
We introduce MARGE, a pre-trained sequence-to-sequence model learned with an unsupervised multi-lingual paraphrasing objective.
We show it is possible to jointly learn to do retrieval and reconstruction, given only a random initialization.
For example, with no additional task-specific training we achieve BLEU scores of up to 35.8 for document translation.
arXiv Detail & Related papers (2020-06-26T14:43:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.