Rubric-Specific Approach to Automated Essay Scoring with Augmentation
Training
- URL: http://arxiv.org/abs/2309.02740v1
- Date: Wed, 6 Sep 2023 05:51:19 GMT
- Title: Rubric-Specific Approach to Automated Essay Scoring with Augmentation
Training
- Authors: Brian Cho, Youngbin Jang, Jaewoong Yoon
- Abstract summary: We propose a series of data augmentation operations that train and test an automated scoring model to learn features and functions overlooked by previous works.
We achieve state-of-the-art performance in the Automated Student Assessment Prize dataset.
- Score: 0.1227734309612871
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Neural based approaches to automatic evaluation of subjective responses have
shown superior performance and efficiency compared to traditional rule-based
and feature engineering oriented solutions. However, it remains unclear whether
the suggested neural solutions are sufficient replacements of human raters as
we find recent works do not properly account for rubric items that are
essential for automated essay scoring during model training and validation. In
this paper, we propose a series of data augmentation operations that train and
test an automated scoring model to learn features and functions overlooked by
previous works while still achieving state-of-the-art performance in the
Automated Student Assessment Prize dataset.
Related papers
- Automatic Die Studies for Ancient Numismatics [3.384989790372139]
Die studies are fundamental to quantifying ancient monetary production.
Few works have attempted to automate this task, and none have been properly released and evaluated from a computer vision perspective.
We propose a fully automatic approach that introduces several innovations compared to previous methods.
arXiv Detail & Related papers (2024-07-30T14:54:54Z) - Combating Missing Modalities in Egocentric Videos at Test Time [92.38662956154256]
Real-world applications often face challenges with incomplete modalities due to privacy concerns, efficiency needs, or hardware issues.
We propose a novel approach to address this issue at test time without requiring retraining.
MiDl represents the first self-supervised, online solution for handling missing modalities exclusively at test time.
arXiv Detail & Related papers (2024-04-23T16:01:33Z) - Augmenting Unsupervised Reinforcement Learning with Self-Reference [63.68018737038331]
Humans possess the ability to draw on past experiences explicitly when learning new tasks.
We propose the Self-Reference (SR) approach, an add-on module explicitly designed to leverage historical information.
Our approach achieves state-of-the-art results in terms of Interquartile Mean (IQM) performance and Optimality Gap reduction on the Unsupervised Reinforcement Learning Benchmark.
arXiv Detail & Related papers (2023-11-16T09:07:34Z) - QualEval: Qualitative Evaluation for Model Improvement [82.73561470966658]
We propose QualEval, which augments quantitative scalar metrics with automated qualitative evaluation as a vehicle for model improvement.
QualEval uses a powerful LLM reasoner and our novel flexible linear programming solver to generate human-readable insights.
We demonstrate that leveraging its insights, for example, improves the absolute performance of the Llama 2 model by up to 15% points relative.
arXiv Detail & Related papers (2023-11-06T00:21:44Z) - From Static Benchmarks to Adaptive Testing: Psychometrics in AI Evaluation [60.14902811624433]
We discuss a paradigm shift from static evaluation methods to adaptive testing.
This involves estimating the characteristics and value of each test item in the benchmark and dynamically adjusting items in real-time.
We analyze the current approaches, advantages, and underlying reasons for adopting psychometrics in AI evaluation.
arXiv Detail & Related papers (2023-06-18T09:54:33Z) - The Effectiveness of a Dynamic Loss Function in Neural Network Based
Automated Essay Scoring [0.0]
We present a dynamic loss function that creates an incentive for the model to predict with the correct distribution, as well as predicting the correct values.
Our loss function achieves this goal without sacrificing any performance achieving a Quadratic Weighted Kappa score of 0.752 on the Automated Student Assessment Prize Automated Essay Scoring dataset.
arXiv Detail & Related papers (2023-05-15T16:39:35Z) - Self-Distillation for Further Pre-training of Transformers [83.84227016847096]
We propose self-distillation as a regularization for a further pre-training stage.
We empirically validate the efficacy of self-distillation on a variety of benchmark datasets for image and text classification tasks.
arXiv Detail & Related papers (2022-09-30T02:25:12Z) - Improving Performance of Automated Essay Scoring by using
back-translation essays and adjusted scores [0.0]
We propose a method to increase the number of essay-score pairs using back-translation and score adjustment.
We evaluate the effectiveness of the augmented data using models from prior work.
The performance of the models was improved by using augmented data to train the models.
arXiv Detail & Related papers (2022-03-01T11:05:43Z) - Exploring the Efficacy of Automatically Generated Counterfactuals for
Sentiment Analysis [17.811597734603144]
We propose an approach to automatically generating counterfactual data for data augmentation and explanation.
A comprehensive evaluation on several different datasets and using a variety of state-of-the-art benchmarks demonstrate how our approach can achieve significant improvements in model performance.
arXiv Detail & Related papers (2021-06-29T10:27:01Z) - On Learning Text Style Transfer with Direct Rewards [101.97136885111037]
Lack of parallel corpora makes it impossible to directly train supervised models for the text style transfer task.
We leverage semantic similarity metrics originally used for fine-tuning neural machine translation models.
Our model provides significant gains in both automatic and human evaluation over strong baselines.
arXiv Detail & Related papers (2020-10-24T04:30:02Z) - Incremental Learning for End-to-End Automatic Speech Recognition [41.297106772785206]
We propose an incremental learning method for end-to-end Automatic Speech Recognition (ASR)
We design a novel explainability-based knowledge distillation for ASR models, which is combined with a response-based knowledge distillation to maintain the original model's predictions and the "reason" for the predictions.
Results on a multi-stage sequential training task show that our method outperforms existing ones in mitigating forgetting.
arXiv Detail & Related papers (2020-05-11T08:18:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.