Modeling the Distribution of Normal Data in Pre-Trained Deep Features
for Anomaly Detection
- URL: http://arxiv.org/abs/2005.14140v2
- Date: Fri, 23 Oct 2020 13:51:36 GMT
- Title: Modeling the Distribution of Normal Data in Pre-Trained Deep Features
for Anomaly Detection
- Authors: Oliver Rippel, Patrick Mertens, Dorit Merhof
- Abstract summary: Anomaly Detection (AD) in images refers to identifying images and image substructures that deviate significantly from the norm.
We show that deep feature representations learned by discriminative models on large natural image datasets are well suited to describe normality.
- Score: 2.9864637081333085
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Anomaly Detection (AD) in images is a fundamental computer vision problem and
refers to identifying images and image substructures that deviate significantly
from the norm. Popular AD algorithms commonly try to learn a model of normality
from scratch using task specific datasets, but are limited to semi-supervised
approaches employing mostly normal data due to the inaccessibility of anomalies
on a large scale combined with the ambiguous nature of anomaly appearance.
We follow an alternative approach and demonstrate that deep feature
representations learned by discriminative models on large natural image
datasets are well suited to describe normality and detect even subtle anomalies
in a transfer learning setting. Our model of normality is established by
fitting a multivariate Gaussian (MVG) to deep feature representations of
classification networks trained on ImageNet using normal data only. By
subsequently applying the Mahalanobis distance as the anomaly score we
outperform the current state of the art on the public MVTec AD dataset,
achieving an AUROC value of $95.8 \pm 1.2$ (mean $\pm$ SEM) over all 15
classes. We further investigate why the learned representations are
discriminative to the AD task using Principal Component Analysis. We find that
the principal components containing little variance in normal data are the ones
crucial for discriminating between normal and anomalous instances. This gives a
possible explanation to the often sub-par performance of AD approaches trained
from scratch using normal data only. By selectively fitting a MVG to these most
relevant components only, we are able to further reduce model complexity while
retaining AD performance. We also investigate setting the working point by
selecting acceptable False Positive Rate thresholds based on the MVG
assumption.
Code available at https://github.com/ORippler/gaussian-ad-mvtec
Related papers
- GLAD: Towards Better Reconstruction with Global and Local Adaptive Diffusion Models for Unsupervised Anomaly Detection [60.78684630040313]
Diffusion models tend to reconstruct normal counterparts of test images with certain noises added.
From the global perspective, the difficulty of reconstructing images with different anomalies is uneven.
We propose a global and local adaptive diffusion model (abbreviated to GLAD) for unsupervised anomaly detection.
arXiv Detail & Related papers (2024-06-11T17:27:23Z) - Anomaly Detection by Context Contrasting [57.695202846009714]
Anomaly detection focuses on identifying samples that deviate from the norm.
Recent advances in self-supervised learning have shown great promise in this regard.
We propose Con$$, which learns through context augmentations.
arXiv Detail & Related papers (2024-05-29T07:59:06Z) - ARC: A Generalist Graph Anomaly Detector with In-Context Learning [62.202323209244]
ARC is a generalist GAD approach that enables a one-for-all'' GAD model to detect anomalies across various graph datasets on-the-fly.
equipped with in-context learning, ARC can directly extract dataset-specific patterns from the target dataset.
Extensive experiments on multiple benchmark datasets from various domains demonstrate the superior anomaly detection performance, efficiency, and generalizability of ARC.
arXiv Detail & Related papers (2024-05-27T02:42:33Z) - Toward Generalist Anomaly Detection via In-context Residual Learning with Few-shot Sample Prompts [25.629973843455495]
Generalist Anomaly Detection (GAD) aims to train one single detection model that can generalize to detect anomalies in diverse datasets from different application domains without further training on the target data.
We introduce a novel approach that learns an in-context residual learning model for GAD, termed InCTRL.
InCTRL is the best performer and significantly outperforms state-of-the-art competing methods.
arXiv Detail & Related papers (2024-03-11T08:07:46Z) - Self-supervised Feature Adaptation for 3D Industrial Anomaly Detection [59.41026558455904]
We focus on multi-modal anomaly detection. Specifically, we investigate early multi-modal approaches that attempted to utilize models pre-trained on large-scale visual datasets.
We propose a Local-to-global Self-supervised Feature Adaptation (LSFA) method to finetune the adaptors and learn task-oriented representation toward anomaly detection.
arXiv Detail & Related papers (2024-01-06T07:30:41Z) - Normality Learning-based Graph Anomaly Detection via Multi-Scale
Contrastive Learning [61.57383634677747]
Graph anomaly detection (GAD) has attracted increasing attention in machine learning and data mining.
Here, we propose a normality learning-based GAD framework via multi-scale contrastive learning networks (NLGAD for abbreviation)
Notably, the proposed algorithm improves the detection performance (up to 5.89% AUC gain) compared with the state-of-the-art methods.
arXiv Detail & Related papers (2023-09-12T08:06:04Z) - CRADL: Contrastive Representations for Unsupervised Anomaly Detection
and Localization [2.8659934481869715]
Unsupervised anomaly detection in medical imaging aims to detect and localize arbitrary anomalies without requiring anomalous data during training.
Most current state-of-the-art methods use latent variable generative models operating directly on the images.
We propose CRADL whose core idea is to model the distribution of normal samples directly in the low-dimensional representation space of an encoder trained with a contrastive pretext-task.
arXiv Detail & Related papers (2023-01-05T16:07:49Z) - Exposing Outlier Exposure: What Can Be Learned From Few, One, and Zero
Outlier Images [26.283734474660484]
We show that specialized AD learning methods seem actually superfluous and huge corpora of data expendable.
We investigate this phenomenon and reveal that one-class methods are more robust towards the particular choice of training outliers.
arXiv Detail & Related papers (2022-05-23T17:23:15Z) - Explainable Deep Few-shot Anomaly Detection with Deviation Networks [123.46611927225963]
We introduce a novel weakly-supervised anomaly detection framework to train detection models.
The proposed approach learns discriminative normality by leveraging the labeled anomalies and a prior probability.
Our model is substantially more sample-efficient and robust, and performs significantly better than state-of-the-art competing methods in both closed-set and open-set settings.
arXiv Detail & Related papers (2021-08-01T14:33:17Z) - Constrained Contrastive Distribution Learning for Unsupervised Anomaly
Detection and Localisation in Medical Images [23.79184121052212]
Unsupervised anomaly detection (UAD) learns one-class classifiers exclusively with normal (i.e., healthy) images.
We propose a novel self-supervised representation learning method, called Constrained Contrastive Distribution learning for anomaly detection (CCD)
Our method outperforms current state-of-the-art UAD approaches on three different colonoscopy and fundus screening datasets.
arXiv Detail & Related papers (2021-03-05T01:56:58Z) - Understanding Anomaly Detection with Deep Invertible Networks through
Hierarchies of Distributions and Features [4.25227087152716]
Convolutional networks learn similar low-level feature distributions when trained on any natural image dataset.
When the discriminative features between inliers and outliers are on a high-level, anomaly detection becomes particularly challenging.
We propose two methods to remove the negative impact of model bias and domain prior on detecting high-level differences.
arXiv Detail & Related papers (2020-06-18T20:56:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.