Do Not Trust a Model Because It is Confident: Uncovering and
Characterizing Unknown Unknowns to Student Success Predictors in Online-Based
Learning
- URL: http://arxiv.org/abs/2212.08532v1
- Date: Fri, 16 Dec 2022 15:32:49 GMT
- Title: Do Not Trust a Model Because It is Confident: Uncovering and
Characterizing Unknown Unknowns to Student Success Predictors in Online-Based
Learning
- Authors: Roberta Galici, Tanja K\"aser, Gianni Fenu, Mirko Marras
- Abstract summary: Student success models might be prone to develop weak spots, i.e., examples hard to accurately classify.
This weakness is one of the main factors undermining users' trust, since model predictions could for instance lead an instructor to not intervene on a student in need.
In this paper, we unveil the need of detecting and characterizing unknown unknowns in student success prediction.
- Score: 10.120425915106727
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Student success models might be prone to develop weak spots, i.e., examples
hard to accurately classify due to insufficient representation during model
creation. This weakness is one of the main factors undermining users' trust,
since model predictions could for instance lead an instructor to not intervene
on a student in need. In this paper, we unveil the need of detecting and
characterizing unknown unknowns in student success prediction in order to
better understand when models may fail. Unknown unknowns include the students
for which the model is highly confident in its predictions, but is actually
wrong. Therefore, we cannot solely rely on the model's confidence when
evaluating the predictions quality. We first introduce a framework for the
identification and characterization of unknown unknowns. We then assess its
informativeness on log data collected from flipped courses and online courses
using quantitative analyses and interviews with instructors. Our results show
that unknown unknowns are a critical issue in this domain and that our
framework can be applied to support their detection. The source code is
available at https://github.com/epfl-ml4ed/unknown-unknowns.
Related papers
- An Ambiguity Measure for Recognizing the Unknowns in Deep Learning [0.0]
We study the understanding of deep neural networks from the scope in which they are trained on.
We propose a measure for quantifying the ambiguity of inputs for any given model.
arXiv Detail & Related papers (2023-12-11T02:57:12Z) - R-Tuning: Instructing Large Language Models to Say `I Don't Know' [66.11375475253007]
Large language models (LLMs) have revolutionized numerous domains with their impressive performance but still face their challenges.
Previous instruction tuning methods force the model to complete a sentence no matter whether the model knows the knowledge or not.
We present a new approach called Refusal-Aware Instruction Tuning (R-Tuning)
Experimental results demonstrate R-Tuning effectively improves a model's ability to answer known questions and refrain from answering unknown questions.
arXiv Detail & Related papers (2023-11-16T08:45:44Z) - Improving the Reliability of Large Language Models by Leveraging
Uncertainty-Aware In-Context Learning [76.98542249776257]
Large-scale language models often face the challenge of "hallucination"
We introduce an uncertainty-aware in-context learning framework to empower the model to enhance or reject its output in response to uncertainty.
arXiv Detail & Related papers (2023-10-07T12:06:53Z) - Trusting Language Models in Education [1.2578554943276923]
We propose to use an XGBoost on top of BERT to output the corrected probabilities.
Our hypothesis is that the level of uncertainty contained in the flow of attention is related to the quality of the model's response itself.
arXiv Detail & Related papers (2023-08-07T18:27:54Z) - Beyond Confidence: Reliable Models Should Also Consider Atypicality [43.012818086415514]
We investigate the relationship between how atypical(rare) a sample or a class is and the reliability of a model's predictions.
We show that predictions for atypical inputs or atypical classes are more overconfident and have lower accuracy.
We propose that models should use not only confidence but also atypicality to improve uncertainty quantification and performance.
arXiv Detail & Related papers (2023-05-29T17:37:09Z) - Reliability-Aware Prediction via Uncertainty Learning for Person Image
Retrieval [51.83967175585896]
UAL aims at providing reliability-aware predictions by considering data uncertainty and model uncertainty simultaneously.
Data uncertainty captures the noise" inherent in the sample, while model uncertainty depicts the model's confidence in the sample's prediction.
arXiv Detail & Related papers (2022-10-24T17:53:20Z) - SF-PATE: Scalable, Fair, and Private Aggregation of Teacher Ensembles [50.90773979394264]
This paper studies a model that protects the privacy of individuals' sensitive information while also allowing it to learn non-discriminatory predictors.
A key characteristic of the proposed model is to enable the adoption of off-the-selves and non-private fair models to create a privacy-preserving and fair model.
arXiv Detail & Related papers (2022-04-11T14:42:54Z) - Explain, Edit, and Understand: Rethinking User Study Design for
Evaluating Model Explanations [97.91630330328815]
We conduct a crowdsourcing study, where participants interact with deception detection models that have been trained to distinguish between genuine and fake hotel reviews.
We observe that for a linear bag-of-words model, participants with access to the feature coefficients during training are able to cause a larger reduction in model confidence in the testing phase when compared to the no-explanation control.
arXiv Detail & Related papers (2021-12-17T18:29:56Z) - A Tale Of Two Long Tails [4.970364068620608]
We identify examples the model is uncertain about and characterize the source of said uncertainty.
We investigate whether the rate of learning in the presence of additional information differs between atypical and noisy examples.
Our results show that well-designed interventions over the course of training can be an effective way to characterize and distinguish between different sources of uncertainty.
arXiv Detail & Related papers (2021-07-27T22:49:59Z) - Plausible Counterfactuals: Auditing Deep Learning Classifiers with
Realistic Adversarial Examples [84.8370546614042]
Black-box nature of Deep Learning models has posed unanswered questions about what they learn from data.
Generative Adversarial Network (GAN) and multi-objectives are used to furnish a plausible attack to the audited model.
Its utility is showcased within a human face classification task, unveiling the enormous potential of the proposed framework.
arXiv Detail & Related papers (2020-03-25T11:08:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.