Diversified Outlier Exposure for Out-of-Distribution Detection via
Informative Extrapolation
- URL: http://arxiv.org/abs/2310.13923v2
- Date: Thu, 26 Oct 2023 06:50:44 GMT
- Title: Diversified Outlier Exposure for Out-of-Distribution Detection via
Informative Extrapolation
- Authors: Jianing Zhu, Geng Yu, Jiangchao Yao, Tongliang Liu, Gang Niu, Masashi
Sugiyama, Bo Han
- Abstract summary: Out-of-distribution (OOD) detection is important for deploying reliable machine learning models on real-world applications.
Recent advances in outlier exposure have shown promising results on OOD detection via fine-tuning model with informatively sampled auxiliary outliers.
We propose a novel framework, namely, Diversified Outlier Exposure (DivOE), for effective OOD detection via informative extrapolation based on the given auxiliary outliers.
- Score: 110.34982764201689
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Out-of-distribution (OOD) detection is important for deploying reliable
machine learning models on real-world applications. Recent advances in outlier
exposure have shown promising results on OOD detection via fine-tuning model
with informatively sampled auxiliary outliers. However, previous methods assume
that the collected outliers can be sufficiently large and representative to
cover the boundary between ID and OOD data, which might be impractical and
challenging. In this work, we propose a novel framework, namely, Diversified
Outlier Exposure (DivOE), for effective OOD detection via informative
extrapolation based on the given auxiliary outliers. Specifically, DivOE
introduces a new learning objective, which diversifies the auxiliary
distribution by explicitly synthesizing more informative outliers for
extrapolation during training. It leverages a multi-step optimization method to
generate novel outliers beyond the original ones, which is compatible with many
variants of outlier exposure. Extensive experiments and analyses have been
conducted to characterize and demonstrate the effectiveness of the proposed
DivOE. The code is publicly available at: https://github.com/tmlr-group/DivOE.
Related papers
- Out-Of-Distribution Detection with Diversification (Provably) [75.44158116183483]
Out-of-distribution (OOD) detection is crucial for ensuring reliable deployment of machine learning models.
Recent advancements focus on utilizing easily accessible auxiliary outliers (e.g., data from the web or other datasets) in training.
We propose a theoretical guarantee, termed Diversity-induced Mixup for OOD detection (diverseMix), which enhances the diversity of auxiliary outlier set for training.
arXiv Detail & Related papers (2024-11-21T11:56:32Z) - What If the Input is Expanded in OOD Detection? [77.37433624869857]
Out-of-distribution (OOD) detection aims to identify OOD inputs from unknown classes.
Various scoring functions are proposed to distinguish it from in-distribution (ID) data.
We introduce a novel perspective, i.e., employing different common corruptions on the input space.
arXiv Detail & Related papers (2024-10-24T06:47:28Z) - Margin-bounded Confidence Scores for Out-of-Distribution Detection [2.373572816573706]
We propose a novel method called Margin bounded Confidence Scores (MaCS) to address the nontrivial OOD detection problem.
MaCS enlarges the disparity between ID and OOD scores, which in turn makes the decision boundary more compact.
Experiments on various benchmark datasets for image classification tasks demonstrate the effectiveness of the proposed method.
arXiv Detail & Related papers (2024-09-22T05:40:25Z) - Learning Multi-Manifold Embedding for Out-Of-Distribution Detection [16.283293167689948]
Out-of-distribution (OOD) samples are crucial for trustworthy AI in real-world applications.
This paper introduces a novel Multi-Manifold Embedding Learning (MMEL) framework for enhanced OOD detection.
MMEL generates representative embeddings and employs a prototype-aware scoring function to differentiate OOD samples.
arXiv Detail & Related papers (2024-09-19T05:43:00Z) - Diffusion based Semantic Outlier Generation via Nuisance Awareness for Out-of-Distribution Detection [9.936136347796413]
Out-of-distribution (OOD) detection has recently shown promising results through training with synthetic OOD datasets.
We propose a novel framework, Semantic Outlier generation via Nuisance Awareness (SONA), which notably produces challenging outliers.
Our approach incorporates SONA guidance, providing separate control over semantic and nuisance regions of ID samples.
arXiv Detail & Related papers (2024-08-27T07:52:44Z) - OAL: Enhancing OOD Detection Using Latent Diffusion [5.357756138014614]
Outlier Aware Learning (OAL) framework synthesizes OOD training data directly in the latent space.
We introduce a mutual information-based contrastive learning approach that amplifies the distinction between In-Distribution (ID) and collected OOD features.
arXiv Detail & Related papers (2024-06-24T11:01:43Z) - Envisioning Outlier Exposure by Large Language Models for Out-of-Distribution Detection [71.93411099797308]
Out-of-distribution (OOD) samples are crucial when deploying machine learning models in open-world scenarios.
We propose to tackle this constraint by leveraging the expert knowledge and reasoning capability of large language models (LLM) to potential Outlier Exposure, termed EOE.
EOE can be generalized to different tasks, including far, near, and fine-language OOD detection.
EOE achieves state-of-the-art performance across different OOD tasks and can be effectively scaled to the ImageNet-1K dataset.
arXiv Detail & Related papers (2024-06-02T17:09:48Z) - EAT: Towards Long-Tailed Out-of-Distribution Detection [55.380390767978554]
This paper addresses the challenging task of long-tailed OOD detection.
The main difficulty lies in distinguishing OOD data from samples belonging to the tail classes.
We propose two simple ideas: (1) Expanding the in-distribution class space by introducing multiple abstention classes, and (2) Augmenting the context-limited tail classes by overlaying images onto the context-rich OOD data.
arXiv Detail & Related papers (2023-12-14T13:47:13Z) - Leveraging Diffusion Disentangled Representations to Mitigate Shortcuts
in Underspecified Visual Tasks [92.32670915472099]
We propose an ensemble diversification framework exploiting the generation of synthetic counterfactuals using Diffusion Probabilistic Models (DPMs)
We show that diffusion-guided diversification can lead models to avert attention from shortcut cues, achieving ensemble diversity performance comparable to previous methods requiring additional data collection.
arXiv Detail & Related papers (2023-10-03T17:37:52Z) - Robust Out-of-Distribution Detection on Deep Probabilistic Generative
Models [0.06372261626436676]
Out-of-distribution (OOD) detection is an important task in machine learning systems.
Deep probabilistic generative models facilitate OOD detection by estimating the likelihood of a data sample.
We propose a new detection metric that operates without outlier exposure.
arXiv Detail & Related papers (2021-06-15T06:36:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.