Fake It Till You Make It: Near-Distribution Novelty Detection by
Score-Based Generative Models
- URL: http://arxiv.org/abs/2205.14297v1
- Date: Sat, 28 May 2022 02:02:53 GMT
- Title: Fake It Till You Make It: Near-Distribution Novelty Detection by
Score-Based Generative Models
- Authors: Hossein Mirzaei, Mohammadreza Salehi, Sajjad Shahabi, Efstratios
Gavves, Cees G. M. Snoek, Mohammad Sabokrou, Mohammad Hossein Rohban
- Abstract summary: existing models either fail or face a dramatic drop under the so-called near-distribution" setting.
We propose to exploit a score-based generative model to produce synthetic near-distribution anomalous data.
Our method improves the near-distribution novelty detection by 6% and passes the state-of-the-art by 1% to 5% across nine novelty detection benchmarks.
- Score: 54.182955830194445
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We aim for image-based novelty detection. Despite considerable progress,
existing models either fail or face a dramatic drop under the so-called
``near-distribution" setting, where the differences between normal and
anomalous samples are subtle. We first demonstrate existing methods experience
up to 20\% decrease in performance in the near-distribution setting. Next, we
propose to exploit a score-based generative model to produce synthetic
near-distribution anomalous data. Our model is then fine-tuned to distinguish
such data from the normal samples. We provide a quantitative as well as
qualitative evaluation of this strategy, and compare the results with a variety
of GAN-based models. Effectiveness of our method for both the near-distribution
and standard novelty detection is assessed through extensive experiments on
datasets in diverse applications such as medical images, object classification,
and quality control. This reveals that our method considerably improves over
existing models, and consistently decreases the gap between the
near-distribution and standard novelty detection performance. Overall, our
method improves the near-distribution novelty detection by 6% and passes the
state-of-the-art by 1% to 5% across nine novelty detection benchmarks. The code
repository is available at https://github.com/rohban-lab/FITYMI
Related papers
- Assessing Sample Quality via the Latent Space of Generative Models [44.59115390303591]
We propose to examine the latent space of a trained generative model to infer generated sample quality.
This is feasible because the quality a generated sample directly relates to the amount of training data resembling it.
We show that the proposed score correlates highly with the sample quality for various generative models including VAEs, GANs and Latent Diffusion Models.
arXiv Detail & Related papers (2024-07-21T14:05:06Z) - Diffusion-based Image Generation for In-distribution Data Augmentation in Surface Defect Detection [8.93281936150572]
We show that diffusion models can be used in industrial scenarios to improve the data augmentation procedure.
We propose a novel approach for data augmentation that mixes out-of-distribution with in-distribution samples.
arXiv Detail & Related papers (2024-06-01T17:09:18Z) - Projection Regret: Reducing Background Bias for Novelty Detection via
Diffusion Models [72.07462371883501]
We propose emphProjection Regret (PR), an efficient novelty detection method that mitigates the bias of non-semantic information.
PR computes the perceptual distance between the test image and its diffusion-based projection to detect abnormality.
Extensive experiments demonstrate that PR outperforms the prior art of generative-model-based novelty detection methods by a significant margin.
arXiv Detail & Related papers (2023-12-05T09:44:47Z) - Variational Classification [51.2541371924591]
We derive a variational objective to train the model, analogous to the evidence lower bound (ELBO) used to train variational auto-encoders.
Treating inputs to the softmax layer as samples of a latent variable, our abstracted perspective reveals a potential inconsistency.
We induce a chosen latent distribution, instead of the implicit assumption found in a standard softmax layer.
arXiv Detail & Related papers (2023-05-17T17:47:19Z) - Improving novelty detection with generative adversarial networks on hand
gesture data [1.3750624267664153]
We propose a novel way of solving the issue of classification of out-of-vocabulary gestures using Artificial Neural Networks (ANNs) trained in the Generative Adversarial Network (GAN) framework.
A generative model augments the data set in an online fashion with new samples and target vectors, while a discriminative model determines the class of the samples.
arXiv Detail & Related papers (2023-04-13T17:50:15Z) - Boosting Out-of-Distribution Detection with Multiple Pre-trained Models [41.66566916581451]
Post hoc detection utilizing pre-trained models has shown promising performance and can be scaled to large-scale problems.
We propose a detection enhancement method by ensembling multiple detection decisions derived from a zoo of pre-trained models.
Our method substantially improves the relative performance by 65.40% and 26.96% on the CIFAR10 and ImageNet benchmarks.
arXiv Detail & Related papers (2022-12-24T12:11:38Z) - Intra-class Adaptive Augmentation with Neighbor Correction for Deep
Metric Learning [99.14132861655223]
We propose a novel intra-class adaptive augmentation (IAA) framework for deep metric learning.
We reasonably estimate intra-class variations for every class and generate adaptive synthetic samples to support hard samples mining.
Our method significantly improves and outperforms the state-of-the-art methods on retrieval performances by 3%-6%.
arXiv Detail & Related papers (2022-11-29T14:52:38Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - Saliency Grafting: Innocuous Attribution-Guided Mixup with Calibrated
Label Mixing [104.630875328668]
Mixup scheme suggests mixing a pair of samples to create an augmented training sample.
We present a novel, yet simple Mixup-variant that captures the best of both worlds.
arXiv Detail & Related papers (2021-12-16T11:27:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.