Layer Adaptive Deep Neural Networks for Out-of-distribution Detection
- URL: http://arxiv.org/abs/2203.00192v1
- Date: Tue, 1 Mar 2022 02:47:33 GMT
- Title: Layer Adaptive Deep Neural Networks for Out-of-distribution Detection
- Authors: Haoliang Wang, Chen Zhao, Xujiang Zhao, Feng Chen
- Abstract summary: We propose a novel layer-adaptive OOD detection framework (LA-OOD) for Deep Neural Networks (DNNs)
We train multiple One-Class SVM OOD detectors simultaneously at the intermediate layers to exploit the full spectrum characteristics encoded at varying depths of DNNs.
LA-OOD is robust against OODs of varying complexity and can outperform state-of-the-art competitors by a large margin on some real-world datasets.
- Score: 14.385491722476036
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: During the forward pass of Deep Neural Networks (DNNs), inputs gradually
transformed from low-level features to high-level conceptual labels. While
features at different layers could summarize the important factors of the
inputs at varying levels, modern out-of-distribution (OOD) detection methods
mostly focus on utilizing their ending layer features. In this paper, we
proposed a novel layer-adaptive OOD detection framework (LA-OOD) for DNNs that
can fully utilize the intermediate layers' outputs. Specifically, instead of
training a unified OOD detector at a fixed ending layer, we train multiple
One-Class SVM OOD detectors simultaneously at the intermediate layers to
exploit the full spectrum characteristics encoded at varying depths of DNNs. We
develop a simple yet effective layer-adaptive policy to identify the best layer
for detecting each potential OOD example. LA-OOD can be applied to any existing
DNNs and does not require access to OOD samples during the training. Using
three DNNs of varying depth and architectures, our experiments demonstrate that
LA-OOD is robust against OODs of varying complexity and can outperform
state-of-the-art competitors by a large margin on some real-world datasets.
Related papers
- Dimensionality-induced information loss of outliers in deep neural networks [29.15751143793406]
Out-of-distribution (OOD) detection is a critical issue for systems using a deep neural network (DNN)
We experimentally clarify this issue by investigating the layer dependence of feature representations from multiple perspectives.
We propose a dimensionality-aware OOD detection method based on alignment of features and weights.
arXiv Detail & Related papers (2024-10-29T01:52:46Z) - Can OOD Object Detectors Learn from Foundation Models? [56.03404530594071]
Out-of-distribution (OOD) object detection is a challenging task due to the absence of open-set OOD data.
Inspired by recent advancements in text-to-image generative models, we study the potential of generative models trained on large-scale open-set data to synthesize OOD samples.
We introduce SyncOOD, a simple data curation method that capitalizes on the capabilities of large foundation models.
arXiv Detail & Related papers (2024-09-08T17:28:22Z) - Rethinking the Evaluation of Out-of-Distribution Detection: A Sorites Paradox [70.57120710151105]
Most existing out-of-distribution (OOD) detection benchmarks classify samples with novel labels as the OOD data.
Some marginal OOD samples actually have close semantic contents to the in-distribution (ID) sample, which makes determining the OOD sample a Sorites Paradox.
We construct a benchmark named Incremental Shift OOD (IS-OOD) to address the issue.
arXiv Detail & Related papers (2024-06-14T09:27:56Z) - GROOD: GRadient-aware Out-Of-Distribution detection in interpolated
manifolds [12.727088216619386]
Out-of-distribution detection in deep neural networks (DNNs) can pose risks in real-world deployments.
We introduce GRadient-aware Out-Of-Distribution detection in.
internative manifold (GROOD), a novel framework that relies on the discriminative power of gradient space.
We show that GROD surpasses the established robustness of state-of-the-art baselines.
arXiv Detail & Related papers (2023-12-22T04:28:43Z) - Pseudo-OOD training for robust language models [78.15712542481859]
OOD detection is a key component of a reliable machine-learning model for any industry-scale application.
We propose POORE - POsthoc pseudo-Ood REgularization, that generates pseudo-OOD samples using in-distribution (IND) data.
We extensively evaluate our framework on three real-world dialogue systems, achieving new state-of-the-art in OOD detection.
arXiv Detail & Related papers (2022-10-17T14:32:02Z) - Igeood: An Information Geometry Approach to Out-of-Distribution
Detection [35.04325145919005]
We introduce Igeood, an effective method for detecting out-of-distribution (OOD) samples.
Igeood applies to any pre-trained neural network, works under various degrees of access to the machine learning model.
We show that Igeood outperforms competing state-of-the-art methods on a variety of network architectures and datasets.
arXiv Detail & Related papers (2022-03-15T11:26:35Z) - Triggering Failures: Out-Of-Distribution detection by learning from
local adversarial attacks in Semantic Segmentation [76.2621758731288]
We tackle the detection of out-of-distribution (OOD) objects in semantic segmentation.
Our main contribution is a new OOD detection architecture called ObsNet associated with a dedicated training scheme based on Local Adversarial Attacks (LAA)
We show it obtains top performances both in speed and accuracy when compared to ten recent methods of the literature on three different datasets.
arXiv Detail & Related papers (2021-08-03T17:09:56Z) - Provably Robust Detection of Out-of-distribution Data (almost) for free [124.14121487542613]
Deep neural networks are known to produce highly overconfident predictions on out-of-distribution (OOD) data.
In this paper we propose a novel method where from first principles we combine a certifiable OOD detector with a standard classifier into an OOD aware classifier.
In this way we achieve the best of two worlds: certifiably adversarially robust OOD detection, even for OOD samples close to the in-distribution, without loss in prediction accuracy and close to state-of-the-art OOD detection performance for non-manipulated OOD data.
arXiv Detail & Related papers (2021-06-08T11:40:49Z) - MOOD: Multi-level Out-of-distribution Detection [13.207044902083057]
Out-of-distribution (OOD) detection is essential to prevent anomalous inputs from causing a model to fail during deployment.
We propose a novel framework, multi-level out-of-distribution detection MOOD, which exploits intermediate classifier outputs for dynamic and efficient OOD inference.
MOOD achieves up to 71.05% computational reduction in inference, while maintaining competitive OOD detection performance.
arXiv Detail & Related papers (2021-04-30T02:18:31Z) - FOOD: Fast Out-Of-Distribution Detector [43.31844129399436]
FOOD is an extended deep neural network (DNN) capable of efficiently detecting OOD samples with minimal inference time overhead.
We evaluate FOOD's detection performance on the SVHN, CIFAR-10, and CIFAR-100 datasets.
Our results demonstrate that in addition to achieving state-of-the-art performance, FOOD is fast and applicable to real-world applications.
arXiv Detail & Related papers (2020-08-16T08:22:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.