An Effective Baseline for Robustness to Distributional Shift
- URL: http://arxiv.org/abs/2105.07107v1
- Date: Sat, 15 May 2021 00:46:11 GMT
- Title: An Effective Baseline for Robustness to Distributional Shift
- Authors: Sunil Thulasidasan, Sushil Thapa, Sayera Dhaubhadel, Gopinath
Chennupati, Tanmoy Bhattacharya, Jeff Bilmes
- Abstract summary: Refraining from confidently predicting when faced with categories of inputs different from those seen during training is an important requirement for the safe deployment of deep learning systems.
We present a simple, but highly effective approach to deal with out-of-distribution detection that uses the principle of abstention.
- Score: 5.627346969563955
- License: http://creativecommons.org/licenses/by-sa/4.0/
- Abstract: Refraining from confidently predicting when faced with categories of inputs
different from those seen during training is an important requirement for the
safe deployment of deep learning systems. While simple to state, this has been
a particularly challenging problem in deep learning, where models often end up
making overconfident predictions in such situations. In this work we present a
simple, but highly effective approach to deal with out-of-distribution
detection that uses the principle of abstention: when encountering a sample
from an unseen class, the desired behavior is to abstain from predicting. Our
approach uses a network with an extra abstention class and is trained on a
dataset that is augmented with an uncurated set that consists of a large number
of out-of-distribution (OoD) samples that are assigned the label of the
abstention class; the model is then trained to learn an effective discriminator
between in and out-of-distribution samples. We compare this relatively simple
approach against a wide variety of more complex methods that have been proposed
both for out-of-distribution detection as well as uncertainty modeling in deep
learning, and empirically demonstrate its effectiveness on a wide variety of of
benchmarks and deep architectures for image recognition and text
classification, often outperforming existing approaches by significant margins.
Given the simplicity and effectiveness of this method, we propose that this
approach be used as a new additional baseline for future work in this domain.
Related papers
- Probably Approximately Precision and Recall Learning [62.912015491907994]
Precision and Recall are foundational metrics in machine learning.
One-sided feedback--where only positive examples are observed during training--is inherent in many practical problems.
We introduce a PAC learning framework where each hypothesis is represented by a graph, with edges indicating positive interactions.
arXiv Detail & Related papers (2024-11-20T04:21:07Z) - When Measures are Unreliable: Imperceptible Adversarial Perturbations
toward Top-$k$ Multi-Label Learning [83.8758881342346]
A novel loss function is devised to generate adversarial perturbations that could achieve both visual and measure imperceptibility.
Experiments on large-scale benchmark datasets demonstrate the superiority of our proposed method in attacking the top-$k$ multi-label systems.
arXiv Detail & Related papers (2023-07-27T13:18:47Z) - Large Class Separation is not what you need for Relational
Reasoning-based OOD Detection [12.578844450586]
Out-Of-Distribution (OOD) detection methods provide a solution by identifying semantic novelty.
Most of these methods leverage a learning stage on the known data, which means training (or fine-tuning) a model to capture the concept of normality.
A viable alternative is that of evaluating similarities in the embedding space produced by large pre-trained models without any further learning effort.
arXiv Detail & Related papers (2023-07-12T14:10:15Z) - On Modality Bias Recognition and Reduction [70.69194431713825]
We study the modality bias problem in the context of multi-modal classification.
We propose a plug-and-play loss function method, whereby the feature space for each label is adaptively learned.
Our method yields remarkable performance improvements compared with the baselines.
arXiv Detail & Related papers (2022-02-25T13:47:09Z) - Contrastive Learning for Fair Representations [50.95604482330149]
Trained classification models can unintentionally lead to biased representations and predictions.
Existing debiasing methods for classification models, such as adversarial training, are often expensive to train and difficult to optimise.
We propose a method for mitigating bias by incorporating contrastive learning, in which instances sharing the same class label are encouraged to have similar representations.
arXiv Detail & Related papers (2021-09-22T10:47:51Z) - A Low Rank Promoting Prior for Unsupervised Contrastive Learning [108.91406719395417]
We construct a novel probabilistic graphical model that effectively incorporates the low rank promoting prior into the framework of contrastive learning.
Our hypothesis explicitly requires that all the samples belonging to the same instance class lie on the same subspace with small dimension.
Empirical evidences show that the proposed algorithm clearly surpasses the state-of-the-art approaches on multiple benchmarks.
arXiv Detail & Related papers (2021-08-05T15:58:25Z) - Learning Diverse Representations for Fast Adaptation to Distribution
Shift [78.83747601814669]
We present a method for learning multiple models, incorporating an objective that pressures each to learn a distinct way to solve the task.
We demonstrate our framework's ability to facilitate rapid adaptation to distribution shift.
arXiv Detail & Related papers (2020-06-12T12:23:50Z) - Simultaneous Perturbation Stochastic Approximation for Few-Shot Learning [0.5801044612920815]
We propose a prototypical-like few-shot learning approach based on the prototypical networks method.
The results of experiments on the benchmark dataset demonstrate that the proposed method is superior to the original networks.
arXiv Detail & Related papers (2020-06-09T09:47:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.