Learning When to Say "I Don't Know"
- URL: http://arxiv.org/abs/2209.04944v1
- Date: Sun, 11 Sep 2022 21:50:03 GMT
- Title: Learning When to Say "I Don't Know"
- Authors: Nicholas Kashani Motlagh, Jim Davis, Tim Anderson, Jeremy Gwinnup
- Abstract summary: We propose a new Reject Option Classification technique to identify and remove regions of uncertainty in the decision space.
We consider an alternative formulation by instead analyzing the complementary reject region and employing a validation set to learn per-class softmax thresholds.
We provide results showing the benefits of the proposed method over na"ively thresholding/uncalibrated softmax scores with 2-D points, imagery, and text classification datasets.
- Score: 0.5505634045241288
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We propose a new Reject Option Classification technique to identify and
remove regions of uncertainty in the decision space for a given neural
classifier and dataset. Such existing formulations employ a learned rejection
(remove)/selection (keep) function and require either a known cost for
rejecting examples or strong constraints on the accuracy or coverage of the
selected examples. We consider an alternative formulation by instead analyzing
the complementary reject region and employing a validation set to learn
per-class softmax thresholds. The goal is to maximize the accuracy of the
selected examples subject to a natural randomness allowance on the rejected
examples (rejecting more incorrect than correct predictions). We provide
results showing the benefits of the proposed method over na\"ively thresholding
calibrated/uncalibrated softmax scores with 2-D points, imagery, and text
classification datasets using state-of-the-art pretrained models. Source code
is available at https://github.com/osu-cvl/learning-idk.
Related papers
- Stacked Confusion Reject Plots (SCORE) [0.7673339435080445]
We argue that common reject curves are too abstract and hard to interpret by non-experts.
We propose Stacked Confusion Reject Plots (SCORE) that offer a more intuitive understanding of the used data.
arXiv Detail & Related papers (2024-06-25T07:59:29Z) - Rejection via Learning Density Ratios [50.91522897152437]
Classification with rejection emerges as a learning paradigm which allows models to abstain from making predictions.
We propose a different distributional perspective, where we seek to find an idealized data distribution which maximizes a pretrained model's performance.
Our framework is tested empirically over clean and noisy datasets.
arXiv Detail & Related papers (2024-05-29T01:32:17Z) - Awareness of uncertainty in classification using a multivariate model and multi-views [1.3048920509133808]
The proposed model regularizes uncertain predictions, and trains to calculate both the predictions and their uncertainty estimations.
Given the multi-view predictions together with their uncertainties and confidences, we proposed several methods to calculate final predictions.
The proposed methodology was tested using CIFAR-10 dataset with clean and noisy labels.
arXiv Detail & Related papers (2024-04-16T06:40:51Z) - Dirichlet-Based Prediction Calibration for Learning with Noisy Labels [40.78497779769083]
Learning with noisy labels can significantly hinder the generalization performance of deep neural networks (DNNs)
Existing approaches address this issue through loss correction or example selection methods.
We propose the textitDirichlet-based Prediction (DPC) method as a solution.
arXiv Detail & Related papers (2024-01-13T12:33:04Z) - XAL: EXplainable Active Learning Makes Classifiers Better Low-resource Learners [71.8257151788923]
We propose a novel Explainable Active Learning framework (XAL) for low-resource text classification.
XAL encourages classifiers to justify their inferences and delve into unlabeled data for which they cannot provide reasonable explanations.
Experiments on six datasets show that XAL achieves consistent improvement over 9 strong baselines.
arXiv Detail & Related papers (2023-10-09T08:07:04Z) - Variational Classification [51.2541371924591]
We derive a variational objective to train the model, analogous to the evidence lower bound (ELBO) used to train variational auto-encoders.
Treating inputs to the softmax layer as samples of a latent variable, our abstracted perspective reveals a potential inconsistency.
We induce a chosen latent distribution, instead of the implicit assumption found in a standard softmax layer.
arXiv Detail & Related papers (2023-05-17T17:47:19Z) - Online Selective Classification with Limited Feedback [82.68009460301585]
We study selective classification in the online learning model, wherein a predictor may abstain from classifying an instance.
Two salient aspects of the setting we consider are that the data may be non-realisable, due to which abstention may be a valid long-term action.
We construct simple versioning-based schemes for any $mu in (0,1],$ that make most $Tmu$ mistakes while incurring smash$tildeO(T1-mu)$ excess abstention against adaptive adversaries.
arXiv Detail & Related papers (2021-10-27T08:00:53Z) - Dash: Semi-Supervised Learning with Dynamic Thresholding [72.74339790209531]
We propose a semi-supervised learning (SSL) approach that uses unlabeled examples to train models.
Our proposed approach, Dash, enjoys its adaptivity in terms of unlabeled data selection.
arXiv Detail & Related papers (2021-09-01T23:52:29Z) - Scalable Marginal Likelihood Estimation for Model Selection in Deep
Learning [78.83598532168256]
Marginal-likelihood based model-selection is rarely used in deep learning due to estimation difficulties.
Our work shows that marginal likelihoods can improve generalization and be useful when validation data is unavailable.
arXiv Detail & Related papers (2021-04-11T09:50:24Z) - Optimal strategies for reject option classifiers [0.0]
In classification with a reject option, the classifier is allowed in uncertain cases to abstain from prediction.
We coin a symmetric definition, the bounded-coverage model, which seeks for a classifier with minimal selective risk and guaranteed coverage.
We propose two algorithms to learn the proper uncertainty score from examples for an arbitrary black-box classifier.
arXiv Detail & Related papers (2021-01-29T11:09:32Z) - Regression with reject option and application to kNN [0.0]
We refer to this framework as regression with reject option as an extension of classification with reject option.
We provide a semi-supervised estimation procedure of the optimal rule involving two datasets.
The resulting predictor with reject option is shown to be almost as good as the optimal predictor with reject option both in terms of risk and rejection rate.
arXiv Detail & Related papers (2020-06-30T08:20:57Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.