Extending the Hint Factory for the assistance dilemma: A novel,
data-driven HelpNeed Predictor for proactive problem-solving help
- URL: http://arxiv.org/abs/2010.04124v1
- Date: Thu, 8 Oct 2020 17:04:03 GMT
- Title: Extending the Hint Factory for the assistance dilemma: A novel,
data-driven HelpNeed Predictor for proactive problem-solving help
- Authors: Mehak Maniktala, Christa Cody, Amy Isvik, Nicholas Lytle, Min Chi,
Tiffany Barnes
- Abstract summary: We present a set of data-driven methods to classify, predict, and prevent unproductive problem-solving steps.
We present a HelpNeed classification, that uses prior student data to determine when students are likely to be unproductive.
We conclude with suggestions on how these HelpNeed methods could be applied in other well-structured open-ended domains.
- Score: 6.188683567894372
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Determining when and whether to provide personalized support is a well-known
challenge called the assistance dilemma. A core problem in solving the
assistance dilemma is the need to discover when students are unproductive so
that the tutor can intervene. Such a task is particularly challenging for
open-ended domains, even those that are well-structured with defined principles
and goals. In this paper, we present a set of data-driven methods to classify,
predict, and prevent unproductive problem-solving steps in the well-structured
open-ended domain of logic. This approach leverages and extends the Hint
Factory, a set of methods that leverages prior student solution attempts to
build data-driven intelligent tutors. We present a HelpNeed classification,
that uses prior student data to determine when students are likely to be
unproductive and need help learning optimal problem-solving strategies. We
present a controlled study to determine the impact of an Adaptive pedagogical
policy that provides proactive hints at the start of each step based on the
outcomes of our HelpNeed predictor: productive vs. unproductive. Our results
show that the students in the Adaptive condition exhibited better training
behaviors, with lower help avoidance, and higher help appropriateness (a higher
chance of receiving help when it was likely to be needed), as measured using
the HelpNeed classifier, when compared to the Control. Furthermore, the results
show that the students who received Adaptive hints based on HelpNeed
predictions during training significantly outperform their Control peers on the
posttest, with the former producing shorter, more optimal solutions in less
time. We conclude with suggestions on how these HelpNeed methods could be
applied in other well-structured open-ended domains.
Related papers
- Learning to Assist Humans without Inferring Rewards [65.28156318196397]
We build upon prior work that studies assistance through the lens of empowerment.
An assistive agent aims to maximize the influence of the human's actions.
We prove that these representations estimate a similar notion of empowerment to that studied by prior work.
arXiv Detail & Related papers (2024-11-04T21:31:04Z) - Learning Task Decomposition to Assist Humans in Competitive Programming [90.4846613669734]
We introduce a novel objective for learning task decomposition, termed value (AssistV)
We collect a dataset of human repair experiences on different decomposed solutions.
Under 177 hours of human study, our method enables non-experts to solve 33.3% more problems, speeds them up by 3.3x, and empowers them to match unassisted experts.
arXiv Detail & Related papers (2024-06-07T03:27:51Z) - Improving Socratic Question Generation using Data Augmentation and Preference Optimization [2.1485350418225244]
Large language models (LLMs) can be used to augment human effort by automatically generating Socratic questions for students.
Existing methods that involve prompting these LLMs sometimes produce invalid outputs.
We propose a data augmentation method to enrich existing Socratic questioning datasets with questions that are invalid in specific ways.
Next, we propose a method to optimize open-source LLMs such as LLama 2 to prefer ground-truth questions over generated invalid ones.
arXiv Detail & Related papers (2024-03-01T00:08:20Z) - Optimising Human-AI Collaboration by Learning Convincing Explanations [62.81395661556852]
We propose a method for a collaborative system that remains safe by having a human making decisions.
Ardent enables efficient and effective decision-making by adapting to individual preferences for explanations.
arXiv Detail & Related papers (2023-11-13T16:00:16Z) - Explainable Data-Driven Optimization: From Context to Decision and Back
Again [76.84947521482631]
Data-driven optimization uses contextual information and machine learning algorithms to find solutions to decision problems with uncertain parameters.
We introduce a counterfactual explanation methodology tailored to explain solutions to data-driven problems.
We demonstrate our approach by explaining key problems in operations management such as inventory management and routing.
arXiv Detail & Related papers (2023-01-24T15:25:16Z) - Enhancing a Student Productivity Model for Adaptive Problem-Solving
Assistance [7.253181280137071]
We present a novel data-driven approach to incorporate students' hint usage in predicting their need for help.
We show empirical evidence to support that such a policy can save students a significant amount of time in training.
We conclude with suggestions on the domains that can benefit from this approach as well as the requirements for adoption.
arXiv Detail & Related papers (2022-07-07T00:41:00Z) - Outcome-Driven Reinforcement Learning via Variational Inference [95.82770132618862]
We discuss a new perspective on reinforcement learning, recasting it as the problem of inferring actions that achieve desired outcomes, rather than a problem of maximizing rewards.
To solve the resulting outcome-directed inference problem, we establish a novel variational inference formulation that allows us to derive a well-shaped reward function.
We empirically demonstrate that this method eliminates the need to design reward functions and leads to effective goal-directed behaviors.
arXiv Detail & Related papers (2021-04-20T18:16:21Z) - Avoiding Help Avoidance: Using Interface Design Changes to Promote
Unsolicited Hint Usage in an Intelligent Tutor [6.639504127104268]
We propose a new hint delivery mechanism called "Assertions" for providing unsolicited hints in a data-driven intelligent tutor.
In systems that only provide hints upon request, hint avoidance results in students not receiving hints when they are needed.
Our results show that Assertions significantly increase unsolicited hint usage compared to Messages.
arXiv Detail & Related papers (2020-09-28T14:39:11Z) - Sequential Transfer in Reinforcement Learning with a Generative Model [48.40219742217783]
We show how to reduce the sample complexity for learning new tasks by transferring knowledge from previously-solved ones.
We derive PAC bounds on its sample complexity which clearly demonstrate the benefits of using this kind of prior knowledge.
We empirically verify our theoretical findings in simple simulated domains.
arXiv Detail & Related papers (2020-07-01T19:53:35Z) - Automatic Discovery of Interpretable Planning Strategies [9.410583483182657]
We introduce AI-Interpret, a method for transforming idiosyncratic policies into simple and interpretable descriptions.
We show that prividing the decision rules generated by AI-Interpret as flowcharts significantly improved people's planning strategies and decisions.
arXiv Detail & Related papers (2020-05-24T12:24:52Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.