Improving unsupervised anomaly localization by applying multi-scale
memories to autoencoders
- URL: http://arxiv.org/abs/2012.11113v1
- Date: Mon, 21 Dec 2020 04:44:40 GMT
- Title: Improving unsupervised anomaly localization by applying multi-scale
memories to autoencoders
- Authors: Yifei Yang, Shibing Xiang, Ruixiang Zhang
- Abstract summary: MMAE.MMAE updates slots at corresponding resolution scale as prototype features during unsupervised learning.
For anomaly detection, we accomplish anomaly removal by replacing the original encoded image features at each scale with most relevant prototype features.
Experimental results on various datasets testify that our MMAE successfully removes anomalies at different scales and performs favorably on several datasets.
- Score: 14.075973859711567
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Autoencoder and its variants have been widely applicated in anomaly
detection.The previous work memory-augmented deep autoencoder proposed
memorizing normality to detect anomaly, however it neglects the feature
discrepancy between different resolution scales, therefore we introduce
multi-scale memories to record scale-specific features and multi-scale
attention fuser between the encoding and decoding module of the autoencoder for
anomaly detection, namely MMAE.MMAE updates slots at corresponding resolution
scale as prototype features during unsupervised learning. For anomaly
detection, we accomplish anomaly removal by replacing the original encoded
image features at each scale with most relevant prototype features,and fuse
these features before feeding to the decoding module to reconstruct image.
Experimental results on various datasets testify that our MMAE successfully
removes anomalies at different scales and performs favorably on several
datasets compared to similar reconstruction-based methods.
Related papers
- A Hierarchically Feature Reconstructed Autoencoder for Unsupervised Anomaly Detection [8.512184778338806]
It consists of a well pre-trained encoder to extract hierarchical feature representations and a decoder to reconstruct these intermediate features from the encoder.
The anomalies can be detected when the decoder fails to reconstruct features well, and then errors of hierarchical feature reconstruction are aggregated into an anomaly map to achieve anomaly localization.
Experiment results show that the proposed method outperforms the state-of-the-art methods on MNIST, Fashion-MNIST, CIFAR-10, and MVTec Anomaly Detection datasets.
arXiv Detail & Related papers (2024-05-15T07:20:27Z) - Feature Purified Transformer With Cross-level Feature Guiding Decoder For Multi-class OOD and Anomaly Deteciton [22.59400036717643]
We introduce the FUTUREG framework, which incorporates two innovative modules: the Feature Purification Module (FPM) and the CFG Decoder.
The FPM constrains the normality boundary within the latent space to effectively filter out anomalous features.
The CFG Decoder uses layer-wise encoder representations to guide the reconstruction of filtered features, preserving fine-grained details.
arXiv Detail & Related papers (2024-04-30T16:45:51Z) - Continuous Memory Representation for Anomaly Detection [24.58611060347548]
CRAD is a novel anomaly detection method for representing normal features within a "continuous" memory.
In an evaluation using the MVTec AD dataset, CRAD significantly outperforms the previous state-of-the-art method by reducing 65.0% of the error for multi-class unified anomaly detection.
arXiv Detail & Related papers (2024-02-28T12:38:44Z) - DiAD: A Diffusion-based Framework for Multi-class Anomaly Detection [55.48770333927732]
We propose a Difusion-based Anomaly Detection (DiAD) framework for multi-class anomaly detection.
It consists of a pixel-space autoencoder, a latent-space Semantic-Guided (SG) network with a connection to the stable diffusion's denoising network, and a feature-space pre-trained feature extractor.
Experiments on MVTec-AD and VisA datasets demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2023-12-11T18:38:28Z) - Open-Vocabulary Video Anomaly Detection [57.552523669351636]
Video anomaly detection (VAD) with weak supervision has achieved remarkable performance in utilizing video-level labels to discriminate whether a video frame is normal or abnormal.
Recent studies attempt to tackle a more realistic setting, open-set VAD, which aims to detect unseen anomalies given seen anomalies and normal videos.
This paper takes a step further and explores open-vocabulary video anomaly detection (OVVAD), in which we aim to leverage pre-trained large models to detect and categorize seen and unseen anomalies.
arXiv Detail & Related papers (2023-11-13T02:54:17Z) - Diversity-Measurable Anomaly Detection [106.07413438216416]
We propose Diversity-Measurable Anomaly Detection (DMAD) framework to enhance reconstruction diversity.
PDM essentially decouples deformation from embedding and makes the final anomaly score more reliable.
arXiv Detail & Related papers (2023-03-09T05:52:42Z) - Self-Supervised Masked Convolutional Transformer Block for Anomaly
Detection [122.4894940892536]
We present a novel self-supervised masked convolutional transformer block (SSMCTB) that comprises the reconstruction-based functionality at a core architectural level.
In this work, we extend our previous self-supervised predictive convolutional attentive block (SSPCAB) with a 3D masked convolutional layer, a transformer for channel-wise attention, as well as a novel self-supervised objective based on Huber loss.
arXiv Detail & Related papers (2022-09-25T04:56:10Z) - AnoViT: Unsupervised Anomaly Detection and Localization with Vision
Transformer-based Encoder-Decoder [3.31490164885582]
We propose a vision transformer-based encoder-decoder model, named AnoViT, to reflect normal information by additionally learning the global relationship between image patches.
The proposed model performed better than the convolution-based model on three benchmark datasets.
arXiv Detail & Related papers (2022-03-21T09:01:37Z) - Anomaly Detection with Prototype-Guided Discriminative Latent Embeddings [29.93983580779689]
We present a novel approach for anomaly detection, which utilizes discriminative prototypes of normal data to reconstruct video frames.
In this way, the model will favor the reconstruction of normal events and distort the reconstruction of abnormal events.
We evaluate the effectiveness of our method on three benchmark datasets and experimental results demonstrate the proposed method outperforms the state-of-the-art.
arXiv Detail & Related papers (2021-04-30T12:16:52Z) - ESAD: End-to-end Deep Semi-supervised Anomaly Detection [85.81138474858197]
We propose a new objective function that measures the KL-divergence between normal and anomalous data.
The proposed method significantly outperforms several state-of-the-arts on multiple benchmark datasets.
arXiv Detail & Related papers (2020-12-09T08:16:35Z) - Unsupervised Anomaly Detection with Adversarial Mirrored AutoEncoders [51.691585766702744]
We propose a variant of Adversarial Autoencoder which uses a mirrored Wasserstein loss in the discriminator to enforce better semantic-level reconstruction.
We put forward an alternative measure of anomaly score to replace the reconstruction-based metric.
Our method outperforms the current state-of-the-art methods for anomaly detection on several OOD detection benchmarks.
arXiv Detail & Related papers (2020-03-24T08:26:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.