Seq2Edits: Sequence Transduction Using Span-level Edit Operations
- URL: http://arxiv.org/abs/2009.11136v1
- Date: Wed, 23 Sep 2020 13:28:38 GMT
- Title: Seq2Edits: Sequence Transduction Using Span-level Edit Operations
- Authors: Felix Stahlberg and Shankar Kumar
- Abstract summary: Seq2Edits is an open-vocabulary approach to sequence editing for natural language processing (NLP) tasks.
We evaluate our method on five NLP tasks (text normalization, sentence fusion, sentence splitting & rephrasing, text simplification, and grammatical error correction)
For grammatical error correction, our method speeds up inference by up to 5.2x compared to full sequence models.
- Score: 10.785577504399077
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose Seq2Edits, an open-vocabulary approach to sequence editing for
natural language processing (NLP) tasks with a high degree of overlap between
input and output texts. In this approach, each sequence-to-sequence
transduction is represented as a sequence of edit operations, where each
operation either replaces an entire source span with target tokens or keeps it
unchanged. We evaluate our method on five NLP tasks (text normalization,
sentence fusion, sentence splitting & rephrasing, text simplification, and
grammatical error correction) and report competitive results across the board.
For grammatical error correction, our method speeds up inference by up to 5.2x
compared to full sequence models because inference time depends on the number
of edits rather than the number of target tokens. For text normalization,
sentence fusion, and grammatical error correction, our approach improves
explainability by associating each edit operation with a human-readable tag.
Related papers
- Reducing Sequence Length by Predicting Edit Operations with Large
Language Models [50.66922361766939]
This paper proposes predicting edit spans for the source text for local sequence transduction tasks.
We apply instruction tuning for Large Language Models on the supervision data of edit spans.
Experiments show that the proposed method achieves comparable performance to the baseline in four tasks.
arXiv Detail & Related papers (2023-05-19T17:51:05Z) - Text Editing as Imitation Game [33.418628166176234]
We reformulate text editing as an imitation game using behavioral cloning.
We introduce a dual decoders structure to parallel the decoding while retaining the dependencies between action tokens.
Our model consistently outperforms the autoregressive baselines in terms of performance, efficiency, and robustness.
arXiv Detail & Related papers (2022-10-21T22:07:04Z) - Text Revision by On-the-Fly Representation Optimization [76.11035270753757]
Current state-of-the-art methods formulate these tasks as sequence-to-sequence learning problems.
We present an iterative in-place editing approach for text revision, which requires no parallel data.
It achieves competitive and even better performance than state-of-the-art supervised methods on text simplification.
arXiv Detail & Related papers (2022-04-15T07:38:08Z) - Reformulating Sentence Ordering as Conditional Text Generation [17.91448517871621]
We present Reorder-BART (RE-BART), a sentence ordering framework.
We reformulate the task as a conditional text-to-marker generation setup.
Our framework achieves the state-of-the-art performance across six datasets in Perfect Match Ratio (PMR) and Kendall's tau ($tau$) metric.
arXiv Detail & Related papers (2021-04-14T18:16:47Z) - Text Editing by Command [82.50904226312451]
A prevailing paradigm in neural text generation is one-shot generation, where text is produced in a single step.
We address this limitation with an interactive text generation setting in which the user interacts with the system by issuing commands to edit existing text.
We show that our Interactive Editor, a transformer-based model trained on this dataset, outperforms baselines and obtains positive results in both automatic and human evaluations.
arXiv Detail & Related papers (2020-10-24T08:00:30Z) - Improving the Efficiency of Grammatical Error Correction with Erroneous
Span Detection and Correction [106.63733511672721]
We propose a novel language-independent approach to improve the efficiency for Grammatical Error Correction (GEC) by dividing the task into two subtasks: Erroneous Span Detection ( ESD) and Erroneous Span Correction (ESC)
ESD identifies grammatically incorrect text spans with an efficient sequence tagging model. ESC leverages a seq2seq model to take the sentence with annotated erroneous spans as input and only outputs the corrected text for these spans.
Experiments show our approach performs comparably to conventional seq2seq approaches in both English and Chinese GEC benchmarks with less than 50% time cost for inference.
arXiv Detail & Related papers (2020-10-07T08:29:11Z) - Incomplete Utterance Rewriting as Semantic Segmentation [57.13577518412252]
We present a novel and extensive approach, which formulates it as a semantic segmentation task.
Instead of generating from scratch, such a formulation introduces edit operations and shapes the problem as prediction of a word-level edit matrix.
Our approach is four times faster than the standard approach in inference.
arXiv Detail & Related papers (2020-09-28T09:29:49Z) - Recurrent Inference in Text Editing [6.4689151804633775]
We propose a new inference method, Recurrence, that iteratively performs editing actions, significantly narrowing the problem space.
In each iteration, encoding the partially edited text, Recurrence decodes the latent representation, generates an action of short, fixed-length, and applies the action to complete a single edit.
For a comprehensive comparison, we introduce three types of text editing tasks: Arithmetic Operators Restoration (AOR), Arithmetic Equation Simplification (AES), Arithmetic Equation Correction (AEC)
arXiv Detail & Related papers (2020-09-26T17:06:29Z) - Fact-aware Sentence Split and Rephrase with Permutation Invariant
Training [93.66323661321113]
Sentence Split and Rephrase aims to break down a complex sentence into several simple sentences with its meaning preserved.
Previous studies tend to address the issue by seq2seq learning from parallel sentence pairs.
We introduce Permutation Training to verifies the effects of order variance in seq2seq learning for this task.
arXiv Detail & Related papers (2020-01-16T07:30:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.