A Novel Data Augmentation Technique for Out-of-Distribution Sample
Detection using Compounded Corruptions
- URL: http://arxiv.org/abs/2207.13916v1
- Date: Thu, 28 Jul 2022 07:17:11 GMT
- Title: A Novel Data Augmentation Technique for Out-of-Distribution Sample
Detection using Compounded Corruptions
- Authors: Ramya S. Hebbalaguppe, Soumya Suvra Goshal, Jatin Prakash, Harshad
Khadilkar, Chetan Arora
- Abstract summary: We propose a novel Compounded Corruption technique for the OOD data augmentation termed CnC.
Unlike current state-of-the-art (SOTA) techniques, CnC does not require backpropagation or ensembling at the test time.
Our extensive comparison with 20 methods from the major conferences in last 4 years show that a model trained using CnC based data augmentation, significantly outperforms SOTA.
- Score: 7.8353348433211165
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Modern deep neural network models are known to erroneously classify
out-of-distribution (OOD) test data into one of the in-distribution (ID)
training classes with high confidence. This can have disastrous consequences
for safety-critical applications. A popular mitigation strategy is to train a
separate classifier that can detect such OOD samples at the test time. In most
practical settings OOD examples are not known at the train time, and hence a
key question is: how to augment the ID data with synthetic OOD samples for
training such an OOD detector? In this paper, we propose a novel Compounded
Corruption technique for the OOD data augmentation termed CnC. One of the major
advantages of CnC is that it does not require any hold-out data apart from the
training set. Further, unlike current state-of-the-art (SOTA) techniques, CnC
does not require backpropagation or ensembling at the test time, making our
method much faster at inference. Our extensive comparison with 20 methods from
the major conferences in last 4 years show that a model trained using CnC based
data augmentation, significantly outperforms SOTA, both in terms of OOD
detection accuracy as well as inference time. We include a detailed post-hoc
analysis to investigate the reasons for the success of our method and identify
higher relative entropy and diversity of CnC samples as probable causes. We
also provide theoretical insights via a piece-wise decomposition analysis on a
two-dimensional dataset to reveal (visually and quantitatively) that our
approach leads to a tighter boundary around ID classes, leading to better
detection of OOD samples. Source code link: https://github.com/cnc-ood
Related papers
- How Does Unlabeled Data Provably Help Out-of-Distribution Detection? [63.41681272937562]
Unlabeled in-the-wild data is non-trivial due to the heterogeneity of both in-distribution (ID) and out-of-distribution (OOD) data.
This paper introduces a new learning framework SAL (Separate And Learn) that offers both strong theoretical guarantees and empirical effectiveness.
arXiv Detail & Related papers (2024-02-05T20:36:33Z) - EAT: Towards Long-Tailed Out-of-Distribution Detection [55.380390767978554]
This paper addresses the challenging task of long-tailed OOD detection.
The main difficulty lies in distinguishing OOD data from samples belonging to the tail classes.
We propose two simple ideas: (1) Expanding the in-distribution class space by introducing multiple abstention classes, and (2) Augmenting the context-limited tail classes by overlaying images onto the context-rich OOD data.
arXiv Detail & Related papers (2023-12-14T13:47:13Z) - Model-free Test Time Adaptation for Out-Of-Distribution Detection [62.49795078366206]
We propose a Non-Parametric Test Time textbfAdaptation framework for textbfDistribution textbfDetection (abbr)
abbr utilizes online test samples for model adaptation during testing, enhancing adaptability to changing data distributions.
We demonstrate the effectiveness of abbr through comprehensive experiments on multiple OOD detection benchmarks.
arXiv Detail & Related papers (2023-11-28T02:00:47Z) - Out-of-Distribution Detection with Hilbert-Schmidt Independence
Optimization [114.43504951058796]
Outlier detection tasks have been playing a critical role in AI safety.
Deep neural network classifiers usually tend to incorrectly classify out-of-distribution (OOD) inputs into in-distribution classes with high confidence.
We propose an alternative probabilistic paradigm that is both practically useful and theoretically viable for the OOD detection tasks.
arXiv Detail & Related papers (2022-09-26T15:59:55Z) - Augmenting Softmax Information for Selective Classification with
Out-of-Distribution Data [7.221206118679026]
We show that existing post-hoc methods perform quite differently compared to when evaluated only on OOD detection.
We propose a novel method for SCOD, Softmax Information Retaining Combination (SIRC), that augments softmax-based confidence scores with feature-agnostic information.
Experiments on a wide variety of ImageNet-scale datasets and convolutional neural network architectures show that SIRC is able to consistently match or outperform the baseline for SCOD.
arXiv Detail & Related papers (2022-07-15T14:39:57Z) - VOS: Learning What You Don't Know by Virtual Outlier Synthesis [23.67449949146439]
Out-of-distribution (OOD) detection has received much attention lately due to its importance in the safe deployment of neural networks.
Previous approaches rely on real outlier datasets for model regularization.
We present VOS, a novel framework for OOD detection by adaptively synthesizing virtual outliers.
arXiv Detail & Related papers (2022-02-02T18:43:01Z) - EARLIN: Early Out-of-Distribution Detection for Resource-efficient
Collaborative Inference [4.826988182025783]
Collaborative inference enables resource-constrained edge devices to make inferences by uploading inputs to a server.
While this setup works cost-effectively for successful inferences, it severely underperforms when the model faces input samples on which the model was not trained.
We propose a novel lightweight OOD detection approach that mines important features from the shallow layers of a pretrained CNN model.
arXiv Detail & Related papers (2021-06-25T18:43:23Z) - Provably Robust Detection of Out-of-distribution Data (almost) for free [124.14121487542613]
Deep neural networks are known to produce highly overconfident predictions on out-of-distribution (OOD) data.
In this paper we propose a novel method where from first principles we combine a certifiable OOD detector with a standard classifier into an OOD aware classifier.
In this way we achieve the best of two worlds: certifiably adversarially robust OOD detection, even for OOD samples close to the in-distribution, without loss in prediction accuracy and close to state-of-the-art OOD detection performance for non-manipulated OOD data.
arXiv Detail & Related papers (2021-06-08T11:40:49Z) - Learn what you can't learn: Regularized Ensembles for Transductive
Out-of-distribution Detection [76.39067237772286]
We show that current out-of-distribution (OOD) detection algorithms for neural networks produce unsatisfactory results in a variety of OOD detection scenarios.
This paper studies how such "hard" OOD scenarios can benefit from adjusting the detection method after observing a batch of the test data.
We propose a novel method that uses an artificial labeling scheme for the test data and regularization to obtain ensembles of models that produce contradictory predictions only on the OOD samples in a test batch.
arXiv Detail & Related papers (2020-12-10T16:55:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.