Learning to Predict Trustworthiness with Steep Slope Loss
- URL: http://arxiv.org/abs/2110.00054v1
- Date: Thu, 30 Sep 2021 19:19:09 GMT
- Title: Learning to Predict Trustworthiness with Steep Slope Loss
- Authors: Yan Luo, Yongkang Wong, Mohan S. Kankanhalli, and Qi Zhao
- Abstract summary: We study the problem of predicting trustworthiness on real-world large-scale datasets.
We observe that the trustworthiness predictors trained with prior-art loss functions are prone to view both correct predictions and incorrect predictions to be trustworthy.
We propose a novel steep slope loss to separate the features w.r.t. correct predictions from the ones w.r.t. incorrect predictions by two slide-like curves that oppose each other.
- Score: 69.40817968905495
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Understanding the trustworthiness of a prediction yielded by a classifier is
critical for the safe and effective use of AI models. Prior efforts have been
proven to be reliable on small-scale datasets. In this work, we study the
problem of predicting trustworthiness on real-world large-scale datasets, where
the task is more challenging due to high-dimensional features, diverse visual
concepts, and large-scale samples. In such a setting, we observe that the
trustworthiness predictors trained with prior-art loss functions, i.e., the
cross entropy loss, focal loss, and true class probability confidence loss, are
prone to view both correct predictions and incorrect predictions to be
trustworthy. The reasons are two-fold. Firstly, correct predictions are
generally dominant over incorrect predictions. Secondly, due to the data
complexity, it is challenging to differentiate the incorrect predictions from
the correct ones on real-world large-scale datasets. To improve the
generalizability of trustworthiness predictors, we propose a novel steep slope
loss to separate the features w.r.t. correct predictions from the ones w.r.t.
incorrect predictions by two slide-like curves that oppose each other. The
proposed loss is evaluated with two representative deep learning models, i.e.,
Vision Transformer and ResNet, as trustworthiness predictors. We conduct
comprehensive experiments and analyses on ImageNet, which show that the
proposed loss effectively improves the generalizability of trustworthiness
predictors. The code and pre-trained trustworthiness predictors for
reproducibility are available at
https://github.com/luoyan407/predict_trustworthiness.
Related papers
- Provably Reliable Conformal Prediction Sets in the Presence of Data Poisoning [53.42244686183879]
Conformal prediction provides model-agnostic and distribution-free uncertainty quantification.
Yet, conformal prediction is not reliable under poisoning attacks where adversaries manipulate both training and calibration data.
We propose reliable prediction sets (RPS): the first efficient method for constructing conformal prediction sets with provable reliability guarantees under poisoning.
arXiv Detail & Related papers (2024-10-13T15:37:11Z) - Error-Driven Uncertainty Aware Training [7.702016079410588]
Error-Driven Uncertainty Aware Training aims to enhance the ability of neural classifiers to estimate their uncertainty correctly.
The EUAT approach operates during the model's training phase by selectively employing two loss functions depending on whether the training examples are correctly or incorrectly predicted.
We evaluate EUAT using diverse neural models and datasets in the image recognition domains considering both non-adversarial and adversarial settings.
arXiv Detail & Related papers (2024-05-02T11:48:14Z) - Cross-Prediction-Powered Inference [15.745692520785074]
Cross-prediction is a method for valid inference powered by machine learning.
We show that cross-prediction is consistently more powerful than an adaptation of prediction-powered inference.
arXiv Detail & Related papers (2023-09-28T17:01:58Z) - Learning Sample Difficulty from Pre-trained Models for Reliable
Prediction [55.77136037458667]
We propose to utilize large-scale pre-trained models to guide downstream model training with sample difficulty-aware entropy regularization.
We simultaneously improve accuracy and uncertainty calibration across challenging benchmarks.
arXiv Detail & Related papers (2023-04-20T07:29:23Z) - Confidence and Dispersity Speak: Characterising Prediction Matrix for
Unsupervised Accuracy Estimation [51.809741427975105]
This work aims to assess how well a model performs under distribution shifts without using labels.
We use the nuclear norm that has been shown to be effective in characterizing both properties.
We show that the nuclear norm is more accurate and robust in accuracy than existing methods.
arXiv Detail & Related papers (2023-02-02T13:30:48Z) - Training Uncertainty-Aware Classifiers with Conformalized Deep Learning [7.837881800517111]
Deep neural networks are powerful tools to detect hidden patterns in data and leverage them to make predictions, but they are not designed to understand uncertainty.
We develop a novel training algorithm that can lead to more dependable uncertainty estimates, without sacrificing predictive power.
arXiv Detail & Related papers (2022-05-12T05:08:10Z) - Learning Uncertainty with Artificial Neural Networks for Improved
Remaining Time Prediction of Business Processes [0.15229257192293202]
This paper is the first to apply these techniques to predictive process monitoring.
We found that they contribute towards more accurate predictions and work quickly.
This leads to many interesting applications, enables an earlier adoption of prediction systems with smaller datasets and fosters a better cooperation with humans.
arXiv Detail & Related papers (2021-05-12T10:18:57Z) - Improving Uncertainty Calibration via Prior Augmented Data [56.88185136509654]
Neural networks have proven successful at learning from complex data distributions by acting as universal function approximators.
They are often overconfident in their predictions, which leads to inaccurate and miscalibrated probabilistic predictions.
We propose a solution by seeking out regions of feature space where the model is unjustifiably overconfident, and conditionally raising the entropy of those predictions towards that of the prior distribution of the labels.
arXiv Detail & Related papers (2021-02-22T07:02:37Z) - Learning Accurate Dense Correspondences and When to Trust Them [161.76275845530964]
We aim to estimate a dense flow field relating two images, coupled with a robust pixel-wise confidence map.
We develop a flexible probabilistic approach that jointly learns the flow prediction and its uncertainty.
Our approach obtains state-of-the-art results on challenging geometric matching and optical flow datasets.
arXiv Detail & Related papers (2021-01-05T18:54:11Z) - Robust Validation: Confident Predictions Even When Distributions Shift [19.327409270934474]
We describe procedures for robust predictive inference, where a model provides uncertainty estimates on its predictions rather than point predictions.
We present a method that produces prediction sets (almost exactly) giving the right coverage level for any test distribution in an $f$-divergence ball around the training population.
An essential component of our methodology is to estimate the amount of expected future data shift and build robustness to it.
arXiv Detail & Related papers (2020-08-10T17:09:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.