Sentence-level Feedback Generation for English Language Learners: Does
Data Augmentation Help?
- URL: http://arxiv.org/abs/2212.08999v1
- Date: Sun, 18 Dec 2022 03:53:44 GMT
- Title: Sentence-level Feedback Generation for English Language Learners: Does
Data Augmentation Help?
- Authors: Shabnam Behzad, Amir Zeldes, Nathan Schneider
- Abstract summary: Given a sentence and an error span, the task is to generate a feedback comment explaining the error.
We experiment with LLMs and also create multiple pseudo datasets for the task, investigating how it affects the performance of our system.
We present our results for the task along with extensive analysis of the generated comments with the aim of aiding future studies in feedback comment generation for English language learners.
- Score: 18.30408619963336
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: In this paper, we present strong baselines for the task of Feedback Comment
Generation for Writing Learning. Given a sentence and an error span, the task
is to generate a feedback comment explaining the error. Sentences and feedback
comments are both in English. We experiment with LLMs and also create multiple
pseudo datasets for the task, investigating how it affects the performance of
our system. We present our results for the task along with extensive analysis
of the generated comments with the aim of aiding future studies in feedback
comment generation for English language learners.
Related papers
- Grammatical Error Feedback: An Implicit Evaluation Approach [32.98100553225724]
Grammatical feedback is crucial for consolidating second language (L2) learning.
Most research in computer-assisted language learning has focused on feedback through grammatical error correction (GEC) systems.
This paper exploits this framework to examine the quality and need for GEC to generate feedback, as well as the system used to generate feedback, using essays from the Cambridge Corpus Learner.
arXiv Detail & Related papers (2024-08-18T18:31:55Z) - Generating Feedback-Ladders for Logical Errors in Programming using Large Language Models [2.1485350418225244]
Large language model (LLM)-based methods have shown great promise in feedback generation for programming assignments.
This paper explores using LLMs to generate a "feedback-ladder", i.e., multiple levels of feedback for the same problem-submission pair.
We evaluate the quality of the generated feedback-ladder via a user study with students, educators, and researchers.
arXiv Detail & Related papers (2024-05-01T03:52:39Z) - Exploring LLM Prompting Strategies for Joint Essay Scoring and Feedback Generation [13.854903594424876]
Large language models (LLMs) have demonstrated strong performance in generating coherent and contextually relevant text.
This work explores several prompting strategies for LLM-based zero-shot and few-shot generation of essay feedback.
Inspired by Chain-of-Thought prompting, we study how and to what extent automated essay scoring (AES) can benefit the quality of generated feedback.
arXiv Detail & Related papers (2024-04-24T12:48:06Z) - Learning from Emotions, Demographic Information and Implicit User
Feedback in Task-Oriented Document-Grounded Dialogues [59.516187851808375]
We introduce FEDI, the first English dialogue dataset for task-oriented document-grounded dialogues annotated with demographic information, user emotions and implicit feedback.
Our experiments with FLAN-T5, GPT-2 and LLaMA-2 show that these data have the potential to improve task completion and the factual consistency of the generated responses and user acceptance.
arXiv Detail & Related papers (2024-01-17T14:52:26Z) - Evaluation of ChatGPT Feedback on ELL Writers' Coherence and Cohesion [0.7028778922533686]
ChatGPT has had a transformative effect on education where students are using it to help with homework assignments and teachers are actively employing it in their teaching practices.
This study evaluated the quality of the feedback generated by ChatGPT regarding the coherence and cohesion of the essays written by English Language learners (ELLs) students.
arXiv Detail & Related papers (2023-10-10T10:25:56Z) - UltraFeedback: Boosting Language Models with Scaled AI Feedback [99.4633351133207]
We present textscUltraFeedback, a large-scale, high-quality, and diversified AI feedback dataset.
Our work validates the effectiveness of scaled AI feedback data in constructing strong open-source chat language models.
arXiv Detail & Related papers (2023-10-02T17:40:01Z) - System-Level Natural Language Feedback [83.24259100437965]
We show how to use feedback to formalize system-level design decisions in a human-in-the-loop-process.
We conduct two case studies of this approach for improving search query and dialog response generation.
We show the combination of system-level and instance-level feedback brings further gains.
arXiv Detail & Related papers (2023-06-23T16:21:40Z) - Training Language Models with Language Feedback at Scale [50.70091340506957]
We introduce learning from Language Feedback (ILF), a new approach that utilizes more informative language feedback.
ILF consists of three steps that are applied iteratively: first, conditioning the language model on the input, an initial LM output, and feedback to generate refinements.
We show theoretically that ILF can be viewed as Bayesian Inference, similar to Reinforcement Learning from human feedback.
arXiv Detail & Related papers (2023-03-28T17:04:15Z) - Error syntax aware augmentation of feedback comment generation dataset [116.73173348201341]
This paper presents a solution to the GenChal 2022 shared task dedicated to feedback comment generation for writing learning.
In terms of this task given a text with an error and a span of the error, a system generates an explanatory note that helps the writer (language learner) to improve their writing skills.
arXiv Detail & Related papers (2022-12-29T12:57:23Z) - How to Probe Sentence Embeddings in Low-Resource Languages: On
Structural Design Choices for Probing Task Evaluation [82.96358326053115]
We investigate sensitivity of probing task results to structural design choices.
We probe embeddings in a multilingual setup with design choices that lie in a'stable region', as we identify for English.
We find that results on English do not transfer to other languages.
arXiv Detail & Related papers (2020-06-16T12:37:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.