Some Grammatical Errors are Frequent, Others are Important
- URL: http://arxiv.org/abs/2205.05730v1
- Date: Wed, 11 May 2022 18:59:20 GMT
- Title: Some Grammatical Errors are Frequent, Others are Important
- Authors: Leshem Choshen, Ofir Shifman, Omri Abend
- Abstract summary: We show that some rare errors are considered disturbing while other common ones are not.
This affects possible directions to improve both systems and their evaluation.
- Score: 32.922128367314194
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In Grammatical Error Correction, systems are evaluated by the number of
errors they correct. However, no one has assessed whether all error types are
equally important. We provide and apply a method to quantify the importance of
different grammatical error types to humans. We show that some rare errors are
considered disturbing while other common ones are not. This affects possible
directions to improve both systems and their evaluation.
Related papers
- An Annotated Dataset of Errors in Premodern Greek and Baselines for Detecting Them [25.238651933622773]
We introduce the first dataset of real errors in premodern Greek.
To create this dataset, we use metrics derived from BERT conditionals to sample 1,000 words more likely to contain errors.
arXiv Detail & Related papers (2024-10-14T20:30:54Z) - A Coin Has Two Sides: A Novel Detector-Corrector Framework for Chinese Spelling Correction [79.52464132360618]
Chinese Spelling Correction (CSC) stands as a foundational Natural Language Processing (NLP) task.
We introduce a novel approach based on error detector-corrector framework.
Our detector is designed to yield two error detection results, each characterized by high precision and recall.
arXiv Detail & Related papers (2024-09-06T09:26:45Z) - GEE! Grammar Error Explanation with Large Language Models [64.16199533560017]
We propose the task of grammar error explanation, where a system needs to provide one-sentence explanations for each grammatical error in a pair of erroneous and corrected sentences.
We analyze the capability of GPT-4 in grammar error explanation, and find that it only produces explanations for 60.2% of the errors using one-shot prompting.
We develop a two-step pipeline that leverages fine-tuned and prompted large language models to perform structured atomic token edit extraction.
arXiv Detail & Related papers (2023-11-16T02:45:47Z) - Rule-Based Error Classification for Analyzing Differences in Frequent
Errors [0.0]
We classify errors for 95,631 code pairs and identify 3.47 errors on average, which are submitted by various levels of programmers on an online judge system.
The analyzed results show that, as for the same introductory problems, errors made by novices are due to the lack of knowledge in programming.
On the other hand, errors made by experts are due to misunderstandings caused by the carelessness of reading problems or the challenges of solving problems differently than usual.
arXiv Detail & Related papers (2023-11-01T13:36:20Z) - Enhancing Grammatical Error Correction Systems with Explanations [45.69642286275681]
Grammatical error correction systems improve written communication by detecting and correcting language mistakes.
We introduce EXPECT, a dataset annotated with evidence words and grammatical error types.
Human evaluation verifies our explainable GEC system's explanations can assist second-language learners in determining whether to accept a correction suggestion.
arXiv Detail & Related papers (2023-05-25T03:00:49Z) - A Syntax-Guided Grammatical Error Correction Model with Dependency Tree
Correction [83.14159143179269]
Grammatical Error Correction (GEC) is a task of detecting and correcting grammatical errors in sentences.
We propose a syntax-guided GEC model (SG-GEC) which adopts the graph attention mechanism to utilize the syntactic knowledge of dependency trees.
We evaluate our model on public benchmarks of GEC task and it achieves competitive results.
arXiv Detail & Related papers (2021-11-05T07:07:48Z) - A Simple and Practical Approach to Improve Misspellings in OCR Text [0.0]
This paper focuses on the identification and correction of non-word errors in OCR text.
Traditional N-gram correction methods can handle single-word errors effectively.
In this paper, we develop an unsupervised method that can handle split and merge errors.
arXiv Detail & Related papers (2021-06-22T19:38:17Z) - Classifying Syntactic Errors in Learner Language [32.93997096837712]
We present a method for classifying syntactic errors in learner language, namely errors whose correction alters the morphosyntactic structure of a sentence.
The methodology builds on the established Universal Dependencies syntactic representation scheme, and provides complementary information to other error-classification systems.
arXiv Detail & Related papers (2020-10-21T14:28:22Z) - On the Robustness of Language Encoders against Grammatical Errors [66.05648604987479]
We collect real grammatical errors from non-native speakers and conduct adversarial attacks to simulate these errors on clean text data.
Results confirm that the performance of all tested models is affected but the degree of impact varies.
arXiv Detail & Related papers (2020-05-12T11:01:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.