Denoising diffusion models for out-of-distribution detection
- URL: http://arxiv.org/abs/2211.07740v4
- Date: Thu, 20 Apr 2023 20:09:54 GMT
- Title: Denoising diffusion models for out-of-distribution detection
- Authors: Mark S. Graham, Walter H.L. Pinaya, Petru-Daniel Tudosiu, Parashkev
Nachev, Sebastien Ourselin, M. Jorge Cardoso
- Abstract summary: We exploit the view of denoising probabilistic diffusion models (DDPM) as denoising autoencoders.
We use DDPMs to reconstruct an input that has been noised to a range of noise levels, and use the resulting multi-dimensional reconstruction error to classify out-of-distribution inputs.
- Score: 2.113925122479677
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Out-of-distribution detection is crucial to the safe deployment of machine
learning systems. Currently, unsupervised out-of-distribution detection is
dominated by generative-based approaches that make use of estimates of the
likelihood or other measurements from a generative model. Reconstruction-based
methods offer an alternative approach, in which a measure of reconstruction
error is used to determine if a sample is out-of-distribution. However,
reconstruction-based approaches are less favoured, as they require careful
tuning of the model's information bottleneck - such as the size of the latent
dimension - to produce good results. In this work, we exploit the view of
denoising diffusion probabilistic models (DDPM) as denoising autoencoders where
the bottleneck is controlled externally, by means of the amount of noise
applied. We propose to use DDPMs to reconstruct an input that has been noised
to a range of noise levels, and use the resulting multi-dimensional
reconstruction error to classify out-of-distribution inputs. We validate our
approach both on standard computer-vision datasets and on higher dimension
medical datasets. Our approach outperforms not only reconstruction-based
methods, but also state-of-the-art generative-based approaches. Code is
available at https://github.com/marksgraham/ddpm-ood.
Related papers
- SEMRes-DDPM: Residual Network Based Diffusion Modelling Applied to
Imbalanced Data [9.969882349165745]
In the field of data mining and machine learning, commonly used classification models cannot effectively learn in unbalanced data.
Most of the classical oversampling methods are based on the SMOTE technique, which only focuses on the local information of the data.
We propose a novel oversampling method SEMRes-DDPM.
arXiv Detail & Related papers (2024-03-09T14:01:04Z) - Classification Diffusion Models: Revitalizing Density Ratio Estimation [21.264497139730473]
$textitclassification diffusion models$ (CDMs) is a DRE-based generative method that adopts the formalism of denoising diffusion models.
Our method is the first DRE-based technique that can successfully generate images beyond the MNIST dataset.
arXiv Detail & Related papers (2024-02-15T16:49:42Z) - Projection Regret: Reducing Background Bias for Novelty Detection via
Diffusion Models [72.07462371883501]
We propose emphProjection Regret (PR), an efficient novelty detection method that mitigates the bias of non-semantic information.
PR computes the perceptual distance between the test image and its diffusion-based projection to detect abnormality.
Extensive experiments demonstrate that PR outperforms the prior art of generative-model-based novelty detection methods by a significant margin.
arXiv Detail & Related papers (2023-12-05T09:44:47Z) - Direct Unsupervised Denoising [60.71146161035649]
Unsupervised denoisers do not directly produce a single prediction, such as the MMSE estimate.
We present an alternative approach that trains a deterministic network alongside the VAE to directly predict a central tendency.
arXiv Detail & Related papers (2023-10-27T13:02:12Z) - AdjointDPM: Adjoint Sensitivity Method for Gradient Backpropagation of Diffusion Probabilistic Models [103.41269503488546]
Existing customization methods require access to multiple reference examples to align pre-trained diffusion probabilistic models with user-provided concepts.
This paper aims to address the challenge of DPM customization when the only available supervision is a differentiable metric defined on the generated contents.
We propose a novel method AdjointDPM, which first generates new samples from diffusion models by solving the corresponding probability-flow ODEs.
It then uses the adjoint sensitivity method to backpropagate the gradients of the loss to the models' parameters.
arXiv Detail & Related papers (2023-07-20T09:06:21Z) - Noise-to-Norm Reconstruction for Industrial Anomaly Detection and
Localization [5.101905755052051]
Anomaly detection has a wide range of applications and is especially important in industrial quality inspection.
Reconstruction-based methods use reconstruction errors to detect anomalies without considering positional differences between samples.
In this study, a reconstruction-based method using the noise-to-norm paradigm is proposed, which avoids the invariant reconstruction of anomalous regions.
arXiv Detail & Related papers (2023-07-06T08:06:48Z) - Improving the Robustness of Summarization Models by Detecting and
Removing Input Noise [50.27105057899601]
We present a large empirical study quantifying the sometimes severe loss in performance from different types of input noise for a range of datasets and model sizes.
We propose a light-weight method for detecting and removing such noise in the input during model inference without requiring any training, auxiliary models, or even prior knowledge of the type of noise.
arXiv Detail & Related papers (2022-12-20T00:33:11Z) - Window-Based Distribution Shift Detection for Deep Neural Networks [21.73028341299301]
We study the case of monitoring the healthy operation of a deep neural network (DNN) receiving a stream of data.
Using selective prediction principles, we propose a distribution deviation detection method for DNNs.
Our novel detection method performs on-par or better than the state-of-the-art, while consuming substantially lower time.
arXiv Detail & Related papers (2022-10-19T21:27:25Z) - Self-Supervised Training with Autoencoders for Visual Anomaly Detection [61.62861063776813]
We focus on a specific use case in anomaly detection where the distribution of normal samples is supported by a lower-dimensional manifold.
We adapt a self-supervised learning regime that exploits discriminative information during training but focuses on the submanifold of normal examples.
We achieve a new state-of-the-art result on the MVTec AD dataset -- a challenging benchmark for visual anomaly detection in the manufacturing domain.
arXiv Detail & Related papers (2022-06-23T14:16:30Z) - Unsupervised Anomaly Detection with Adversarial Mirrored AutoEncoders [51.691585766702744]
We propose a variant of Adversarial Autoencoder which uses a mirrored Wasserstein loss in the discriminator to enforce better semantic-level reconstruction.
We put forward an alternative measure of anomaly score to replace the reconstruction-based metric.
Our method outperforms the current state-of-the-art methods for anomaly detection on several OOD detection benchmarks.
arXiv Detail & Related papers (2020-03-24T08:26:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.