COLA-GEC: A Bidirectional Framework for Enhancing Grammatical Acceptability and Error Correction
- URL: http://arxiv.org/abs/2507.11867v1
- Date: Wed, 16 Jul 2025 03:29:05 GMT
- Title: COLA-GEC: A Bidirectional Framework for Enhancing Grammatical Acceptability and Error Correction
- Authors: Xiangyu Yang, Xinying Qiu,
- Abstract summary: This paper introduces COLA-GEC, a novel bidirectional framework that enhances both tasks through mutual knowledge transfer.<n>First, we augment grammatical acceptability models using GEC datasets, significantly improving their performance across multiple languages.<n>Second, we integrate grammatical acceptability signals into GEC model training via a dynamic loss function, effectively guiding corrections toward grammatically acceptable outputs.
- Score: 2.631955426232593
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Grammatical Error Correction (GEC) and grammatical acceptability judgment (COLA) are core tasks in natural language processing, sharing foundational grammatical knowledge yet typically evolving independently. This paper introduces COLA-GEC, a novel bidirectional framework that enhances both tasks through mutual knowledge transfer. First, we augment grammatical acceptability models using GEC datasets, significantly improving their performance across multiple languages. Second, we integrate grammatical acceptability signals into GEC model training via a dynamic loss function, effectively guiding corrections toward grammatically acceptable outputs. Our approach achieves state-of-the-art results on several multilingual benchmarks. Comprehensive error analysis highlights remaining challenges, particularly in punctuation error correction, providing insights for future improvements in grammatical modeling.
Related papers
- Corrections Meet Explanations: A Unified Framework for Explainable Grammatical Error Correction [29.583603444317855]
We introduce EXGEC, a unified explainable GEC framework that integrates explanation and correction tasks in a generative manner.<n>Results on various NLP models (BART, T5, and Llama3) show that EXGEC models surpass single-task baselines in both tasks.
arXiv Detail & Related papers (2025-02-21T07:42:33Z) - Detection-Correction Structure via General Language Model for Grammatical Error Correction [22.609760120265587]
This paper introduces an integrated detection-correction structure, named DeCoGLM, based on the General Language Model (GLM)
The detection phase employs a fault-tolerant detection template, while the correction phase leverages autoregressive mask infilling for localized error correction.
Our model demonstrates competitive performance against the state-of-the-art models on English and Chinese GEC datasets.
arXiv Detail & Related papers (2024-05-28T04:04:40Z) - Grammatical Error Correction via Mixed-Grained Weighted Training [68.94921674855621]
Grammatical Error Correction (GEC) aims to automatically correct grammatical errors in natural texts.
MainGEC designs token-level and sentence-level training weights based on inherent discrepancies in accuracy and potential diversity of data annotation.
arXiv Detail & Related papers (2023-11-23T08:34:37Z) - Generative error correction for code-switching speech recognition using
large language models [49.06203730433107]
Code-switching (CS) speech refers to the phenomenon of mixing two or more languages within the same sentence.
We propose to leverage large language models (LLMs) and lists of hypotheses generated by an ASR to address the CS problem.
arXiv Detail & Related papers (2023-10-17T14:49:48Z) - Enhancing Grammatical Error Correction Systems with Explanations [45.69642286275681]
Grammatical error correction systems improve written communication by detecting and correcting language mistakes.
We introduce EXPECT, a dataset annotated with evidence words and grammatical error types.
Human evaluation verifies our explainable GEC system's explanations can assist second-language learners in determining whether to accept a correction suggestion.
arXiv Detail & Related papers (2023-05-25T03:00:49Z) - A Unified Strategy for Multilingual Grammatical Error Correction with
Pre-trained Cross-Lingual Language Model [100.67378875773495]
We propose a generic and language-independent strategy for multilingual Grammatical Error Correction.
Our approach creates diverse parallel GEC data without any language-specific operations.
It achieves the state-of-the-art results on the NLPCC 2018 Task 2 dataset (Chinese) and obtains competitive performance on Falko-Merlin (German) and RULEC-GEC (Russian)
arXiv Detail & Related papers (2022-01-26T02:10:32Z) - A Syntax-Guided Grammatical Error Correction Model with Dependency Tree
Correction [83.14159143179269]
Grammatical Error Correction (GEC) is a task of detecting and correcting grammatical errors in sentences.
We propose a syntax-guided GEC model (SG-GEC) which adopts the graph attention mechanism to utilize the syntactic knowledge of dependency trees.
We evaluate our model on public benchmarks of GEC task and it achieves competitive results.
arXiv Detail & Related papers (2021-11-05T07:07:48Z) - Improving the Efficiency of Grammatical Error Correction with Erroneous
Span Detection and Correction [106.63733511672721]
We propose a novel language-independent approach to improve the efficiency for Grammatical Error Correction (GEC) by dividing the task into two subtasks: Erroneous Span Detection ( ESD) and Erroneous Span Correction (ESC)
ESD identifies grammatically incorrect text spans with an efficient sequence tagging model. ESC leverages a seq2seq model to take the sentence with annotated erroneous spans as input and only outputs the corrected text for these spans.
Experiments show our approach performs comparably to conventional seq2seq approaches in both English and Chinese GEC benchmarks with less than 50% time cost for inference.
arXiv Detail & Related papers (2020-10-07T08:29:11Z) - Towards Minimal Supervision BERT-based Grammar Error Correction [81.90356787324481]
We try to incorporate contextual information from pre-trained language model to leverage annotation and benefit multilingual scenarios.
Results show strong potential of Bidirectional Representations from Transformers (BERT) in grammatical error correction task.
arXiv Detail & Related papers (2020-01-10T15:45:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.