Mahalanobis-Aware Training for Out-of-Distribution Detection
- URL: http://arxiv.org/abs/2311.00808v1
- Date: Wed, 1 Nov 2023 19:46:40 GMT
- Title: Mahalanobis-Aware Training for Out-of-Distribution Detection
- Authors: Connor Mclaughlin, Jason Matterer, Michael Yee
- Abstract summary: We present a novel loss function and recipe for training networks with improved density-based out-of-distribution sensitivity.
We demonstrate the effectiveness of our method on CIFAR-10, notably reducing the false-positive rate of the relative Mahalanobis distance method on far-OOD tasks by over 50%.
- Score: 0.11510009152620666
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: While deep learning models have seen widespread success in controlled
environments, there are still barriers to their adoption in open-world
settings. One critical task for safe deployment is the detection of anomalous
or out-of-distribution samples that may require human intervention. In this
work, we present a novel loss function and recipe for training networks with
improved density-based out-of-distribution sensitivity. We demonstrate the
effectiveness of our method on CIFAR-10, notably reducing the false-positive
rate of the relative Mahalanobis distance method on far-OOD tasks by over 50%.
Related papers
- Out-of-Distribution Detection in Medical Imaging via Diffusion Trajectories [8.591748099943167]
In medical imaging, unsupervised out-of-distribution (OOD) detection offers an attractive approach for identifying pathological cases.<n>Current generative approaches often rely on likelihood estimation or reconstruction error.<n>We propose a reconstruction-free OOD detection method that leverages the forward diffusion trajectories of a Stein score-based denoising diffusion model.
arXiv Detail & Related papers (2025-07-31T10:36:58Z) - Leveraging Perturbation Robustness to Enhance Out-of-Distribution Detection [15.184096796229115]
We propose a post-hoc method, Perturbation-Rectified OOD detection (PRO), based on the insight that prediction confidence for OOD inputs is more susceptible to reduction under perturbation than in-distribution (IND) inputs.
On a CIFAR-10 model with adversarial training, PRO effectively detects near-OOD inputs, achieving a reduction of more than 10% on FPR@95 compared to state-of-the-art methods.
arXiv Detail & Related papers (2025-03-24T15:32:33Z) - Lie Detector: Unified Backdoor Detection via Cross-Examination Framework [68.45399098884364]
We propose a unified backdoor detection framework in the semi-honest setting.
Our method achieves superior detection performance, improving accuracy by 5.4%, 1.6%, and 11.9% over SoTA baselines.
Notably, it is the first to effectively detect backdoors in multimodal large language models.
arXiv Detail & Related papers (2025-03-21T06:12:06Z) - Embracing Unknown Step by Step: Towards Reliable Sparse Training in Real World [23.836541532002126]
Sparse training has emerged as a promising method for resource-efficient deep neural networks (DNNs) in real-world applications.
However, the reliability of sparse models remains a crucial concern, particularly in detecting unknown out-of-distribution (OOD) data.
This study addresses the knowledge gap by investigating the reliability of sparse training from an OOD perspective.
We propose a new unknown-aware sparse training method, which incorporates a loss modification, auto-tuning strategy, and a voting scheme.
arXiv Detail & Related papers (2024-03-29T08:33:05Z) - T2FNorm: Extremely Simple Scaled Train-time Feature Normalization for
OOD Detection [16.03174062601543]
We introduce T2FNorm, a novel approach to transforming features to hyperspherical space during training, while employing non-transformed space for OOD-scoring purposes.
This method yields a surprising enhancement in OOD detection capabilities without compromising model accuracy in in-distribution(ID)
arXiv Detail & Related papers (2023-05-28T18:56:54Z) - Average of Pruning: Improving Performance and Stability of
Out-of-Distribution Detection [37.43981354073841]
We find the performance of OOD detection suffers from overfitting and instability during training.
We propose Average of Pruning (AoP), consisting of model averaging and pruning, to mitigate the unstable behaviors.
arXiv Detail & Related papers (2023-03-02T12:34:38Z) - Anti-Exploration by Random Network Distillation [63.04360288089277]
We show that a naive choice of conditioning for the Random Network Distillation (RND) is not discriminative enough to be used as an uncertainty estimator.
We show that this limitation can be avoided with conditioning based on Feature-wise Linear Modulation (FiLM)
We evaluate it on the D4RL benchmark, showing that it is capable of achieving performance comparable to ensemble-based methods and outperforming ensemble-free approaches by a wide margin.
arXiv Detail & Related papers (2023-01-31T13:18:33Z) - Diffusion Denoising Process for Perceptron Bias in Out-of-distribution
Detection [67.49587673594276]
We introduce a new perceptron bias assumption that suggests discriminator models are more sensitive to certain features of the input, leading to the overconfidence problem.
We demonstrate that the diffusion denoising process (DDP) of DMs serves as a novel form of asymmetric, which is well-suited to enhance the input and mitigate the overconfidence problem.
Our experiments on CIFAR10, CIFAR100, and ImageNet show that our method outperforms SOTA approaches.
arXiv Detail & Related papers (2022-11-21T08:45:08Z) - Improving Out-of-Distribution Detection via Epistemic Uncertainty
Adversarial Training [29.4569172720654]
We develop a simple adversarial training scheme that incorporates an attack of the uncertainty predicted by the dropout ensemble.
We demonstrate this method improves OOD detection performance on standard data (i.e., not adversarially crafted), and improves the standardized partial AUC from near-random guessing performance to $geq 0.75$.
arXiv Detail & Related papers (2022-09-05T14:32:19Z) - Controlled Sparsity via Constrained Optimization or: How I Learned to
Stop Tuning Penalties and Love Constraints [81.46143788046892]
We focus on the task of controlling the level of sparsity when performing sparse learning.
Existing methods based on sparsity-inducing penalties involve expensive trial-and-error tuning of the penalty factor.
We propose a constrained formulation where sparsification is guided by the training objective and the desired sparsity target in an end-to-end fashion.
arXiv Detail & Related papers (2022-08-08T21:24:20Z) - To be Critical: Self-Calibrated Weakly Supervised Learning for Salient
Object Detection [95.21700830273221]
Weakly-supervised salient object detection (WSOD) aims to develop saliency models using image-level annotations.
We propose a self-calibrated training strategy by explicitly establishing a mutual calibration loop between pseudo labels and network predictions.
We prove that even a much smaller dataset with well-matched annotations can facilitate models to achieve better performance as well as generalizability.
arXiv Detail & Related papers (2021-09-04T02:45:22Z) - On the Practicality of Deterministic Epistemic Uncertainty [106.06571981780591]
deterministic uncertainty methods (DUMs) achieve strong performance on detecting out-of-distribution data.
It remains unclear whether DUMs are well calibrated and can seamlessly scale to real-world applications.
arXiv Detail & Related papers (2021-07-01T17:59:07Z) - Robust Out-of-distribution Detection for Neural Networks [51.19164318924997]
We show that existing detection mechanisms can be extremely brittle when evaluating on in-distribution and OOD inputs.
We propose an effective algorithm called ALOE, which performs robust training by exposing the model to both adversarially crafted inlier and outlier examples.
arXiv Detail & Related papers (2020-03-21T17:46:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.