Effective Feedback for Introductory CS Theory: A JFLAP Extension and
Student Persistence
- URL: http://arxiv.org/abs/2012.01546v1
- Date: Wed, 2 Dec 2020 21:39:01 GMT
- Title: Effective Feedback for Introductory CS Theory: A JFLAP Extension and
Student Persistence
- Authors: Ivona Bez\'akov\'a, Kimberly Fluet, Edith Hemaspaandra, Hannah Miller,
David E. Narv\'aez
- Abstract summary: The main goal of our research is to help students learn abstract computational models.
The most common pedagogical tool for interacting with these models is the Java Formal Languages and Automata Package (JFLAP)
We developed a JFLAP server extension, which accepts homework submissions from students, evaluates the submission as correct or incorrect, and provides a witness string when the submission is incorrect.
- Score: 4.40401067183266
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Computing theory analyzes abstract computational models to rigorously study
the computational difficulty of various problems. Introductory computing theory
can be challenging for undergraduate students, and the main goal of our
research is to help students learn these computational models. The most common
pedagogical tool for interacting with these models is the Java Formal Languages
and Automata Package (JFLAP). We developed a JFLAP server extension, which
accepts homework submissions from students, evaluates the submission as correct
or incorrect, and provides a witness string when the submission is incorrect.
Our extension currently provides witness feedback for deterministic finite
automata, nondeterministic finite automata, regular expressions, context-free
grammars, and pushdown automata.
In Fall 2019, we ran a preliminary investigation on two sections (Control and
Study) of the required undergraduate course Introduction to Computer Science
Theory. The Study section used our extension for five targeted homework
questions, and the Control section solved and submitted these problems using
traditional means. Our results show that on these five questions, the Study
section performed better on average than the Control section. Moreover, the
Study section persisted in submitting attempts until correct, and from this
finding, our preliminary conclusion is that minimal (not detailed or
grade-based) witness feedback helps students to truly learn the concepts. We
describe the results that support this conclusion as well as a related
hypothesis conjecturing that with witness feedback and unlimited number of
submissions, partial credit is both unnecessary and ineffective.
Related papers
- ReasonAgain: Using Extractable Symbolic Programs to Evaluate Mathematical Reasoning [54.70811660561151]
Existing math datasets evaluate the reasoning abilities of large language models (LLMs) by either using the final answer or the intermediate reasoning steps derived from static examples.
We seek to use symbolic programs as a means for automated evaluation if a model can consistently produce correct final answers across various inputs to the program.
We observe significant accuracy drops using our proposed evaluation compared with original static examples, suggesting the fragility of math reasoning in state-of-the-art LLMs.
arXiv Detail & Related papers (2024-10-24T18:02:37Z) - Investigating Video Reasoning Capability of Large Language Models with Tropes in Movies [69.28082193942991]
This paper introduces a novel dataset, Tropes in Movies (TiM), designed as a testbed for exploring two critical yet previously overlooked video reasoning skills.
utilizing tropes from movie storytelling, TiM evaluates the reasoning capabilities of state-of-the-art LLM-based approaches.
To address these deficiencies, we propose Face-Enhanced Viper of Role Interactions (FEVoRI) and Context Query Reduction (ConQueR)
arXiv Detail & Related papers (2024-06-16T12:58:31Z) - The potential of large language models for improving probability
learning: A study on ChatGPT3.5 and first-year computer engineering students [0.565395466029518]
ChatGPT is a large-scale language model that can solve probability problems.
ChatGPT is used in solving probability problems typically presented in computer engineering exams.
The model's ability to deliver high-quality explanations and illustrate solutions in any programming language suggests that large language models have the potential to serve as learning assistants.
arXiv Detail & Related papers (2023-10-09T12:54:58Z) - Exploring the Potential of Large Language Models to Generate Formative
Programming Feedback [0.5371337604556311]
We explore the potential of large language models (LLMs) for computing educators and learners.
To achieve these goals, we used students' programming sequences from a dataset gathered within a CS1 course as input for ChatGPT.
Results show that ChatGPT performs reasonably well for some of the introductory programming tasks and student errors.
However, educators should provide guidance on how to use the provided feedback, as it can contain misleading information for novices.
arXiv Detail & Related papers (2023-08-31T15:22:11Z) - Investigating Fairness Disparities in Peer Review: A Language Model
Enhanced Approach [77.61131357420201]
We conduct a thorough and rigorous study on fairness disparities in peer review with the help of large language models (LMs)
We collect, assemble, and maintain a comprehensive relational database for the International Conference on Learning Representations (ICLR) conference from 2017 to date.
We postulate and study fairness disparities on multiple protective attributes of interest, including author gender, geography, author, and institutional prestige.
arXiv Detail & Related papers (2022-11-07T16:19:42Z) - From Human Days to Machine Seconds: Automatically Answering and
Generating Machine Learning Final Exams [10.25071232250652]
A final exam in machine learning at a top institution such as MIT, Harvard, or Cornell typically takes faculty days to write, and students hours to solve.
We demonstrate that large language models pass machine learning finals at a human level, on finals available online after the models were trained, and automatically generate new human-quality final exam questions in seconds.
arXiv Detail & Related papers (2022-06-11T06:38:06Z) - Automatic Short Math Answer Grading via In-context Meta-learning [2.0263791972068628]
We study the problem of automatic short answer grading for students' responses to math questions.
We use MathBERT, a variant of the popular language model BERT adapted to mathematical content, as our base model.
Second, we use an in-context learning approach that provides scoring examples as input to the language model.
arXiv Detail & Related papers (2022-05-30T16:26:02Z) - Textual Explanations and Critiques in Recommendation Systems [8.406549970145846]
dissertation focuses on two fundamental challenges of addressing this need.
The first involves explanation generation in a scalable and data-driven manner.
The second challenge consists in making explanations actionable, and we refer to it as critiquing.
arXiv Detail & Related papers (2022-05-15T11:59:23Z) - Unifying Language Learning Paradigms [96.35981503087567]
We present a unified framework for pre-training models that are universally effective across datasets and setups.
We show how different pre-training objectives can be cast as one another and how interpolating between different objectives can be effective.
Our model also achieve strong results at in-context learning, outperforming 175B GPT-3 on zero-shot SuperGLUE and tripling the performance of T5-XXL on one-shot summarization.
arXiv Detail & Related papers (2022-05-10T19:32:20Z) - ProtoTransformer: A Meta-Learning Approach to Providing Student Feedback [54.142719510638614]
In this paper, we frame the problem of providing feedback as few-shot classification.
A meta-learner adapts to give feedback to student code on a new programming question from just a few examples by instructors.
Our approach was successfully deployed to deliver feedback to 16,000 student exam-solutions in a programming course offered by a tier 1 university.
arXiv Detail & Related papers (2021-07-23T22:41:28Z) - Exploring Bayesian Deep Learning for Urgent Instructor Intervention Need
in MOOC Forums [58.221459787471254]
Massive Open Online Courses (MOOCs) have become a popular choice for e-learning thanks to their great flexibility.
Due to large numbers of learners and their diverse backgrounds, it is taxing to offer real-time support.
With the large volume of posts and high workloads for MOOC instructors, it is unlikely that the instructors can identify all learners requiring intervention.
This paper explores for the first time Bayesian deep learning on learner-based text posts with two methods: Monte Carlo Dropout and Variational Inference.
arXiv Detail & Related papers (2021-04-26T15:12:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.