Sequence-to-Action: Grammatical Error Correction with Action Guided
Sequence Generation
- URL: http://arxiv.org/abs/2205.10884v1
- Date: Sun, 22 May 2022 17:47:06 GMT
- Title: Sequence-to-Action: Grammatical Error Correction with Action Guided
Sequence Generation
- Authors: Jiquan Li, Junliang Guo, Yongxin Zhu, Xin Sheng, Deqiang Jiang, Bo
Ren, Linli Xu
- Abstract summary: We propose a novel Sequence-to-Action(S2A) module for Grammatical Error Correction.
The S2A module jointly takes the source and target sentences as input, and is able to automatically generate a token-level action sequence.
Our model consistently outperforms the seq2seq baselines, while being able to significantly alleviate the over-correction problem.
- Score: 21.886973310718457
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The task of Grammatical Error Correction (GEC) has received remarkable
attention with wide applications in Natural Language Processing (NLP) in recent
years. While one of the key principles of GEC is to keep the correct parts
unchanged and avoid over-correction, previous sequence-to-sequence (seq2seq)
models generate results from scratch, which are not guaranteed to follow the
original sentence structure and may suffer from the over-correction problem. In
the meantime, the recently proposed sequence tagging models can overcome the
over-correction problem by only generating edit operations, but are conditioned
on human designed language-specific tagging labels. In this paper, we combine
the pros and alleviate the cons of both models by proposing a novel
Sequence-to-Action~(S2A) module. The S2A module jointly takes the source and
target sentences as input, and is able to automatically generate a token-level
action sequence before predicting each token, where each action is generated
from three choices named SKIP, COPY and GENerate. Then the actions are fused
with the basic seq2seq framework to provide final predictions. We conduct
experiments on the benchmark datasets of both English and Chinese GEC tasks.
Our model consistently outperforms the seq2seq baselines, while being able to
significantly alleviate the over-correction problem as well as holding better
generality and diversity in the generation results compared to the sequence
tagging models.
Related papers
- Multi-head Sequence Tagging Model for Grammatical Error Correction [31.538895931875565]
Grammatical Error Correction (GEC) problem is a mapping between a source sequence and a target one.
Current sequence tagging approaches still have some issues handling a broad range of grammatical errors just by being laser-focused on one task.
We propose a novel multi-head and multi-task learning model to effectively utilize training data and harness the information from related task training signals.
arXiv Detail & Related papers (2024-10-21T20:01:06Z) - Improving Seq2Seq Grammatical Error Correction via Decoding
Interventions [40.52259641181596]
We propose a unified decoding intervention framework that employs an external critic to assess the appropriateness of the token to be generated incrementally.
We discover and investigate two types of critics: a pre-trained left-to-right language model critic and an incremental target-side grammatical error detector critic.
Our framework consistently outperforms strong baselines and achieves results competitive with state-of-the-art methods.
arXiv Detail & Related papers (2023-10-23T03:36:37Z) - SequenceMatch: Imitation Learning for Autoregressive Sequence Modelling with Backtracking [60.109453252858806]
A maximum-likelihood (MLE) objective does not match a downstream use-case of autoregressively generating high-quality sequences.
We formulate sequence generation as an imitation learning (IL) problem.
This allows us to minimize a variety of divergences between the distribution of sequences generated by an autoregressive model and sequences from a dataset.
Our resulting method, SequenceMatch, can be implemented without adversarial training or architectural changes.
arXiv Detail & Related papers (2023-06-08T17:59:58Z) - Mutual Exclusivity Training and Primitive Augmentation to Induce
Compositionality [84.94877848357896]
Recent datasets expose the lack of the systematic generalization ability in standard sequence-to-sequence models.
We analyze this behavior of seq2seq models and identify two contributing factors: a lack of mutual exclusivity bias and the tendency to memorize whole examples.
We show substantial empirical improvements using standard sequence-to-sequence models on two widely-used compositionality datasets.
arXiv Detail & Related papers (2022-11-28T17:36:41Z) - Hierarchical Phrase-based Sequence-to-Sequence Learning [94.10257313923478]
We describe a neural transducer that maintains the flexibility of standard sequence-to-sequence (seq2seq) models while incorporating hierarchical phrases as a source of inductive bias during training and as explicit constraints during inference.
Our approach trains two models: a discriminative derivation based on a bracketing grammar whose tree hierarchically aligns source and target phrases, and a neural seq2seq model that learns to translate the aligned phrases one-by-one.
arXiv Detail & Related papers (2022-11-15T05:22:40Z) - Don't Take It Literally: An Edit-Invariant Sequence Loss for Text
Generation [109.46348908829697]
We propose a novel Edit-Invariant Sequence Loss (EISL), which computes the matching loss of a target n-gram with all n-grams in the generated sequence.
We conduct experiments on three tasks: machine translation with noisy target sequences, unsupervised text style transfer, and non-autoregressive machine translation.
arXiv Detail & Related papers (2021-06-29T03:59:21Z) - Structured Reordering for Modeling Latent Alignments in Sequence
Transduction [86.94309120789396]
We present an efficient dynamic programming algorithm performing exact marginal inference of separable permutations.
The resulting seq2seq model exhibits better systematic generalization than standard models on synthetic problems and NLP tasks.
arXiv Detail & Related papers (2021-06-06T21:53:54Z) - Tail-to-Tail Non-Autoregressive Sequence Prediction for Chinese
Grammatical Error Correction [49.25830718574892]
We present a new framework named Tail-to-Tail (textbfTtT) non-autoregressive sequence prediction.
Considering that most tokens are correct and can be conveyed directly from source to target, and the error positions can be estimated and corrected.
Experimental results on standard datasets, especially on the variable-length datasets, demonstrate the effectiveness of TtT in terms of sentence-level Accuracy, Precision, Recall, and F1-Measure.
arXiv Detail & Related papers (2021-06-03T05:56:57Z) - Grammatical Error Correction as GAN-like Sequence Labeling [45.19453732703053]
We propose a GAN-like sequence labeling model, which consists of a grammatical error detector as a discriminator and a grammatical error labeler with Gumbel-Softmax sampling as a generator.
Our results on several evaluation benchmarks demonstrate that our proposed approach is effective and improves the previous state-of-the-art baseline.
arXiv Detail & Related papers (2021-05-29T04:39:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.