Neural Quality Estimation with Multiple Hypotheses for Grammatical Error
Correction
- URL: http://arxiv.org/abs/2105.04443v1
- Date: Mon, 10 May 2021 15:04:25 GMT
- Title: Neural Quality Estimation with Multiple Hypotheses for Grammatical Error
Correction
- Authors: Zhenghao Liu, Xiaoyuan Yi, Maosong Sun, Liner Yang and Tat-Seng Chua
- Abstract summary: Grammatical Error Correction (GEC) aims to correct writing errors and help language learners improve their writing skills.
Existing GEC models tend to produce spurious corrections or fail to detect lots of errors.
This paper presents the Neural Verification Network (VERNet) for GEC quality estimation with multiple hypotheses.
- Score: 98.31440090585376
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Grammatical Error Correction (GEC) aims to correct writing errors and help
language learners improve their writing skills. However, existing GEC models
tend to produce spurious corrections or fail to detect lots of errors. The
quality estimation model is necessary to ensure learners get accurate GEC
results and avoid misleading from poorly corrected sentences. Well-trained GEC
models can generate several high-quality hypotheses through decoding, such as
beam search, which provide valuable GEC evidence and can be used to evaluate
GEC quality. However, existing models neglect the possible GEC evidence from
different hypotheses. This paper presents the Neural Verification Network
(VERNet) for GEC quality estimation with multiple hypotheses. VERNet
establishes interactions among hypotheses with a reasoning graph and conducts
two kinds of attention mechanisms to propagate GEC evidence to verify the
quality of generated hypotheses. Our experiments on four GEC datasets show that
VERNet achieves state-of-the-art grammatical error detection performance,
achieves the best quality estimation results, and significantly improves GEC
performance by reranking hypotheses. All data and source codes are available at
https://github.com/thunlp/VERNet.
Related papers
- Failing Forward: Improving Generative Error Correction for ASR with Synthetic Data and Retrieval Augmentation [73.9145653659403]
We show that Generative Error Correction models struggle to generalize beyond the specific types of errors encountered during training.
We propose DARAG, a novel approach designed to improve GEC for ASR in in-domain (ID) and OOD scenarios.
Our approach is simple, scalable, and both domain- and language-agnostic.
arXiv Detail & Related papers (2024-10-17T04:00:29Z) - Online GNN Evaluation Under Test-time Graph Distribution Shifts [92.4376834462224]
A new research problem, online GNN evaluation, aims to provide valuable insights into the well-trained GNNs's ability to generalize to real-world unlabeled graphs.
We develop an effective learning behavior discrepancy score, dubbed LeBeD, to estimate the test-time generalization errors of well-trained GNN models.
arXiv Detail & Related papers (2024-03-15T01:28:08Z) - System Combination via Quality Estimation for Grammatical Error
Correction [29.91720235173108]
We propose GRECO, a new state-of-the-art quality estimation model that gives a better estimate of the quality of a corrected sentence.
We also propose three methods for utilizing GEC quality estimation models for system combination with varying generality.
arXiv Detail & Related papers (2023-10-23T13:46:49Z) - RobustGEC: Robust Grammatical Error Correction Against Subtle Context
Perturbation [64.2568239429946]
We introduce RobustGEC, a benchmark designed to evaluate the context robustness of GEC systems.
We reveal that state-of-the-art GEC systems still lack sufficient robustness against context perturbations.
arXiv Detail & Related papers (2023-10-11T08:33:23Z) - FCGEC: Fine-Grained Corpus for Chinese Grammatical Error Correction [6.116341682577877]
Grammatical Error Correction (GEC) has been broadly applied in automatic correction and proofreading system recently.
We present FCGEC, a fine-grained corpus to detect, identify and correct the grammatical errors.
arXiv Detail & Related papers (2022-10-22T06:29:05Z) - Proficiency Matters Quality Estimation in Grammatical Error Correction [30.31557952622774]
This study investigates how supervised quality estimation (QE) models of grammatical error correction (GEC) are affected by the learners' proficiency with the data.
arXiv Detail & Related papers (2022-01-17T03:47:19Z) - ErAConD : Error Annotated Conversational Dialog Dataset for Grammatical
Error Correction [30.917993017459615]
We present a novel parallel grammatical error correction (GEC) dataset drawn from open-domain conversations.
This dataset is, to our knowledge, the first GEC dataset targeted to a conversational setting.
To demonstrate the utility of the dataset, we use our annotated data to fine-tune a state-of-the-art GEC model.
arXiv Detail & Related papers (2021-12-15T20:27:40Z) - A Syntax-Guided Grammatical Error Correction Model with Dependency Tree
Correction [83.14159143179269]
Grammatical Error Correction (GEC) is a task of detecting and correcting grammatical errors in sentences.
We propose a syntax-guided GEC model (SG-GEC) which adopts the graph attention mechanism to utilize the syntactic knowledge of dependency trees.
We evaluate our model on public benchmarks of GEC task and it achieves competitive results.
arXiv Detail & Related papers (2021-11-05T07:07:48Z) - A Self-Refinement Strategy for Noise Reduction in Grammatical Error
Correction [54.569707226277735]
Existing approaches for grammatical error correction (GEC) rely on supervised learning with manually created GEC datasets.
There is a non-negligible amount of "noise" where errors were inappropriately edited or left uncorrected.
We propose a self-refinement method where the key idea is to denoise these datasets by leveraging the prediction consistency of existing models.
arXiv Detail & Related papers (2020-10-07T04:45:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.