Bangla Grammatical Error Detection Using T5 Transformer Model
- URL: http://arxiv.org/abs/2303.10612v1
- Date: Sun, 19 Mar 2023 09:24:48 GMT
- Title: Bangla Grammatical Error Detection Using T5 Transformer Model
- Authors: H.A.Z. Sameen Shahgir, Khondker Salman Sayeed
- Abstract summary: This paper presents a method for detecting grammatical errors in Bangla using a Text-to-Text Transfer Transformer (T5 Language Model)
The T5 model was primarily designed for translation and is not specifically designed for this task, so extensive post-processing was necessary to adapt it to the task of error detection.
Our experiments show that the T5 model can achieve low Levenshtein Distance in detecting grammatical errors in Bangla, but post-processing is essential to achieve optimal performance.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: This paper presents a method for detecting grammatical errors in Bangla using
a Text-to-Text Transfer Transformer (T5) Language Model, using the small
variant of BanglaT5, fine-tuned on a corpus of 9385 sentences where errors were
bracketed by the dedicated demarcation symbol. The T5 model was primarily
designed for translation and is not specifically designed for this task, so
extensive post-processing was necessary to adapt it to the task of error
detection. Our experiments show that the T5 model can achieve low Levenshtein
Distance in detecting grammatical errors in Bangla, but post-processing is
essential to achieve optimal performance. The final average Levenshtein
Distance after post-processing the output of the fine-tuned model was 1.0394 on
a test set of 5000 sentences. This paper also presents a detailed analysis of
the errors detected by the model and discusses the challenges of adapting a
translation model for grammar. Our approach can be extended to other languages,
demonstrating the potential of T5 models for detecting grammatical errors in a
wide range of languages.
Related papers
- Bangla Grammatical Error Detection Leveraging Transformer-based Token Classification [0.0]
We study the development of an automated grammar checker in Bangla, the seventh most spoken language in the world.
Our approach involves breaking down the task as a token classification problem and utilizing state-of-the-art transformer-based models.
Our system is evaluated on a dataset consisting of over 25,000 texts from various sources.
arXiv Detail & Related papers (2024-11-13T05:22:45Z) - mmT5: Modular Multilingual Pre-Training Solves Source Language
Hallucinations [54.42422445568523]
mmT5 is a modular multilingual sequence-to-sequence model.
It disentangles language-specific information from language-agnostic information.
Compared to mT5, mmT5 raises the rate of generating text in the correct language under zero-shot settings from 7% to 99%.
arXiv Detail & Related papers (2023-05-23T16:38:01Z) - Towards Fine-Grained Information: Identifying the Type and Location of
Translation Errors [80.22825549235556]
Existing approaches can not synchronously consider error position and type.
We build an FG-TED model to predict the textbf addition and textbfomission errors.
Experiments show that our model can identify both error type and position concurrently, and gives state-of-the-art results.
arXiv Detail & Related papers (2023-02-17T16:20:33Z) - idT5: Indonesian Version of Multilingual T5 Transformer [0.0]
Indonesian is spoken by almost 200 million people and is the 10th most spoken language in the world.
In this study, the mT5 model was adapted for only one language, Indonesian, resulting in a pre-trained T5 model that was specific only for Indonesian with a smaller size.
Fine-tuned model based on our model achieved 77.18% accuracy on SA, 8% higher than the mT5-based model, and obtained nearly the same score as the mT5-based model on QG and QA.
arXiv Detail & Related papers (2023-02-02T03:56:16Z) - Evaluating Byte and Wordpiece Level Models for Massively Multilingual
Semantic Parsing [3.431659287330068]
We compare a byte-level (ByT5) and a wordpiece based (mT5) sequence to sequence model on the 51 languages of the MASSIVE multilingual semantic parsing dataset.
We are able to reduce the gap in exact match accuracy to only 5 points with respect to a model trained on gold data from all the languages.
arXiv Detail & Related papers (2022-12-14T13:48:32Z) - Conciseness: An Overlooked Language Task [11.940413163824887]
We define the task and show that it is different from related tasks such as summarization and simplification.
We demonstrate that conciseness is a difficult task for which zero-shot setups with large neural language models often do not perform well.
arXiv Detail & Related papers (2022-11-08T09:47:11Z) - mT5: A massively multilingual pre-trained text-to-text transformer [60.0210636815514]
"Text-to-Text Transfer Transformer" (T5) leveraged a unified text-to-text format and scale to attain state-of-the-art results on English-language NLP tasks.
We introduce mT5, a multilingual variant of T5 that was pre-trained on a new Common Crawl-based dataset covering 101 languages.
arXiv Detail & Related papers (2020-10-22T17:58:14Z) - On the Robustness of Language Encoders against Grammatical Errors [66.05648604987479]
We collect real grammatical errors from non-native speakers and conduct adversarial attacks to simulate these errors on clean text data.
Results confirm that the performance of all tested models is affected but the degree of impact varies.
arXiv Detail & Related papers (2020-05-12T11:01:44Z) - TuringAdvice: A Generative and Dynamic Evaluation of Language Use [90.3029315711237]
We propose TuringAdvice, a new challenge task and dataset for language understanding models.
Given a written situation that a real person is currently facing, a model must generate helpful advice in natural language.
Empirical results show that today's models struggle at TuringAdvice.
arXiv Detail & Related papers (2020-04-07T18:00:03Z) - Towards Minimal Supervision BERT-based Grammar Error Correction [81.90356787324481]
We try to incorporate contextual information from pre-trained language model to leverage annotation and benefit multilingual scenarios.
Results show strong potential of Bidirectional Representations from Transformers (BERT) in grammatical error correction task.
arXiv Detail & Related papers (2020-01-10T15:45:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.