On the Impact of Spurious Correlation for Out-of-distribution Detection
- URL: http://arxiv.org/abs/2109.05642v1
- Date: Sun, 12 Sep 2021 23:58:17 GMT
- Title: On the Impact of Spurious Correlation for Out-of-distribution Detection
- Authors: Yifei Ming, Hang Yin, Yixuan Li
- Abstract summary: We present a new formalization and model the data shifts by taking into account both the invariant and environmental features.
Our results suggest that the detection performance is severely worsened when the correlation between spurious features and labels is increased in the training set.
- Score: 14.186776881154127
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Modern neural networks can assign high confidence to inputs drawn from
outside the training distribution, posing threats to models in real-world
deployments. While much research attention has been placed on designing new
out-of-distribution (OOD) detection methods, the precise definition of OOD is
often left in vagueness and falls short of the desired notion of OOD in
reality. In this paper, we present a new formalization and model the data
shifts by taking into account both the invariant and environmental (spurious)
features. Under such formalization, we systematically investigate how spurious
correlation in the training set impacts OOD detection. Our results suggest that
the detection performance is severely worsened when the correlation between
spurious features and labels is increased in the training set. We further show
insights on detection methods that are more effective in reducing the impact of
spurious correlation and provide theoretical analysis on why reliance on
environmental features leads to high OOD detection error. Our work aims to
facilitate a better understanding of OOD samples and their formalization, as
well as the exploration of methods that enhance OOD detection.
Related papers
- Double Descent Meets Out-of-Distribution Detection: Theoretical Insights and Empirical Analysis on the role of model complexity [2.206582444513284]
We propose an expected OOD risk metric to evaluate classifiers confidence on both training and OOD samples.
We show that the OOD risk depicts an infinite peak, when the number of parameters is equal to the number of samples.
arXiv Detail & Related papers (2024-11-04T15:39:12Z) - NECO: NEural Collapse Based Out-of-distribution detection [2.4958897155282282]
We introduce NECO, a novel post-hoc method for OOD detection.
Our experiments demonstrate that NECO achieves both small and large-scale OOD detection tasks.
We provide a theoretical explanation for the effectiveness of our method in OOD detection.
arXiv Detail & Related papers (2023-10-10T17:53:36Z) - Meta OOD Learning for Continuously Adaptive OOD Detection [38.28089655572316]
Out-of-distribution (OOD) detection is crucial to modern deep learning applications.
We propose a novel and more realistic setting called continuously adaptive out-of-distribution (CAOOD) detection.
We develop meta OOD learning (MOL) by designing a learning-to-adapt diagram such that a good OOD detection model is learned during the training process.
arXiv Detail & Related papers (2023-09-21T01:05:45Z) - Beyond AUROC & co. for evaluating out-of-distribution detection
performance [50.88341818412508]
Given their relevance for safe(r) AI, it is important to examine whether the basis for comparing OOD detection methods is consistent with practical needs.
We propose a new metric - Area Under the Threshold Curve (AUTC), which explicitly penalizes poor separation between ID and OOD samples.
arXiv Detail & Related papers (2023-06-26T12:51:32Z) - LINe: Out-of-Distribution Detection by Leveraging Important Neurons [15.797257361788812]
We introduce a new aspect for analyzing the difference in model outputs between in-distribution data and OOD data.
We propose a novel method, Leveraging Important Neurons (LINe), for post-hoc Out of distribution detection.
arXiv Detail & Related papers (2023-03-24T13:49:05Z) - Out-of-distribution Detection with Implicit Outlier Transformation [72.73711947366377]
Outlier exposure (OE) is powerful in out-of-distribution (OOD) detection.
We propose a novel OE-based approach that makes the model perform well for unseen OOD situations.
arXiv Detail & Related papers (2023-03-09T04:36:38Z) - Energy-based Out-of-Distribution Detection for Graph Neural Networks [76.0242218180483]
We propose a simple, powerful and efficient OOD detection model for GNN-based learning on graphs, which we call GNNSafe.
GNNSafe achieves up to $17.0%$ AUROC improvement over state-of-the-arts and it could serve as simple yet strong baselines in such an under-developed area.
arXiv Detail & Related papers (2023-02-06T16:38:43Z) - Pseudo-OOD training for robust language models [78.15712542481859]
OOD detection is a key component of a reliable machine-learning model for any industry-scale application.
We propose POORE - POsthoc pseudo-Ood REgularization, that generates pseudo-OOD samples using in-distribution (IND) data.
We extensively evaluate our framework on three real-world dialogue systems, achieving new state-of-the-art in OOD detection.
arXiv Detail & Related papers (2022-10-17T14:32:02Z) - Learn what you can't learn: Regularized Ensembles for Transductive
Out-of-distribution Detection [76.39067237772286]
We show that current out-of-distribution (OOD) detection algorithms for neural networks produce unsatisfactory results in a variety of OOD detection scenarios.
This paper studies how such "hard" OOD scenarios can benefit from adjusting the detection method after observing a batch of the test data.
We propose a novel method that uses an artificial labeling scheme for the test data and regularization to obtain ensembles of models that produce contradictory predictions only on the OOD samples in a test batch.
arXiv Detail & Related papers (2020-12-10T16:55:13Z) - Robust Out-of-distribution Detection for Neural Networks [51.19164318924997]
We show that existing detection mechanisms can be extremely brittle when evaluating on in-distribution and OOD inputs.
We propose an effective algorithm called ALOE, which performs robust training by exposing the model to both adversarially crafted inlier and outlier examples.
arXiv Detail & Related papers (2020-03-21T17:46:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.