DFR: Deep Feature Reconstruction for Unsupervised Anomaly Segmentation
- URL: http://arxiv.org/abs/2012.07122v1
- Date: Sun, 13 Dec 2020 18:30:51 GMT
- Title: DFR: Deep Feature Reconstruction for Unsupervised Anomaly Segmentation
- Authors: Jie Yang, Yong Shi, Zhiquan Qi
- Abstract summary: This paper proposes an effective unsupervised anomaly segmentation approach.
It can detect and segment out the anomalies in small and confined regions of images.
It advances the state-of-the-art performances on several benchmark datasets.
- Score: 24.52418722578279
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Automatic detecting anomalous regions in images of objects or textures
without priors of the anomalies is challenging, especially when the anomalies
appear in very small areas of the images, making difficult-to-detect visual
variations, such as defects on manufacturing products. This paper proposes an
effective unsupervised anomaly segmentation approach that can detect and
segment out the anomalies in small and confined regions of images. Concretely,
we develop a multi-scale regional feature generator that can generate multiple
spatial context-aware representations from pre-trained deep convolutional
networks for every subregion of an image. The regional representations not only
describe the local characteristics of corresponding regions but also encode
their multiple spatial context information, making them discriminative and very
beneficial for anomaly detection. Leveraging these descriptive regional
features, we then design a deep yet efficient convolutional autoencoder and
detect anomalous regions within images via fast feature reconstruction. Our
method is simple yet effective and efficient. It advances the state-of-the-art
performances on several benchmark datasets and shows great potential for real
applications.
Related papers
- Weakly Supervised Video Anomaly Detection and Localization with Spatio-Temporal Prompts [57.01985221057047]
This paper introduces a novel method that learnstemporal prompt embeddings for weakly supervised video anomaly detection and localization (WSVADL) based on pre-trained vision-language models (VLMs)
Our method achieves state-of-theart performance on three public benchmarks for the WSVADL task.
arXiv Detail & Related papers (2024-08-12T03:31:29Z) - DiAD: A Diffusion-based Framework for Multi-class Anomaly Detection [55.48770333927732]
We propose a Difusion-based Anomaly Detection (DiAD) framework for multi-class anomaly detection.
It consists of a pixel-space autoencoder, a latent-space Semantic-Guided (SG) network with a connection to the stable diffusion's denoising network, and a feature-space pre-trained feature extractor.
Experiments on MVTec-AD and VisA datasets demonstrate the effectiveness of our approach.
arXiv Detail & Related papers (2023-12-11T18:38:28Z) - DQnet: Cross-Model Detail Querying for Camouflaged Object Detection [54.82390534024954]
A convolutional neural network (CNN) for camouflaged object detection tends to activate local discriminative regions while ignoring complete object extent.
In this paper, we argue that partial activation is caused by the intrinsic characteristics of CNN.
In order to obtain feature maps that could activate full object extent, a novel framework termed Cross-Model Detail Querying network (DQnet) is proposed.
arXiv Detail & Related papers (2022-12-16T06:23:58Z) - Prototypical Residual Networks for Anomaly Detection and Localization [80.5730594002466]
We propose a framework called Prototypical Residual Network (PRN)
PRN learns feature residuals of varying scales and sizes between anomalous and normal patterns to accurately reconstruct the segmentation maps of anomalous regions.
We present a variety of anomaly generation strategies that consider both seen and unseen appearance variance to enlarge and diversify anomalies.
arXiv Detail & Related papers (2022-12-05T05:03:46Z) - Cross-Domain Local Characteristic Enhanced Deepfake Video Detection [18.430287055542315]
Deepfake detection has attracted increasing attention due to security concerns.
Many detectors cannot achieve accurate results when detecting unseen manipulations.
We propose a novel pipeline, Cross-Domain Local Forensics, for more general deepfake video detection.
arXiv Detail & Related papers (2022-11-07T07:44:09Z) - PAEDID: Patch Autoencoder Based Deep Image Decomposition For Pixel-level
Defective Region Segmentation [16.519583839906904]
Unsupervised patch autoencoder based deep image decomposition (PAEDID) method for defective region segmentation is presented.
We learn the common background as a deep image prior by a patch autoencoder (PAE) network.
By adopting the proposed approach, the defective regions in the image can be accurately extracted in an unsupervised fashion.
arXiv Detail & Related papers (2022-03-28T02:50:06Z) - AnoViT: Unsupervised Anomaly Detection and Localization with Vision
Transformer-based Encoder-Decoder [3.31490164885582]
We propose a vision transformer-based encoder-decoder model, named AnoViT, to reflect normal information by additionally learning the global relationship between image patches.
The proposed model performed better than the convolution-based model on three benchmark datasets.
arXiv Detail & Related papers (2022-03-21T09:01:37Z) - Low-Rank Subspaces in GANs [101.48350547067628]
This work introduces low-rank subspaces that enable more precise control of GAN generation.
LowRankGAN is able to find the low-dimensional representation of attribute manifold.
Experiments on state-of-the-art GAN models (including StyleGAN2 and BigGAN) trained on various datasets demonstrate the effectiveness of our LowRankGAN.
arXiv Detail & Related papers (2021-06-08T16:16:32Z) - Anomaly localization by modeling perceptual features [3.04585143845864]
Feature-Augmented VAE is trained by reconstructing the input image in pixel space, and also in several different feature spaces.
It achieves clear improvement over state-of-the-art methods on the MVTec anomaly detection and localization datasets.
arXiv Detail & Related papers (2020-08-12T15:09:13Z) - Image Fine-grained Inpainting [89.17316318927621]
We present a one-stage model that utilizes dense combinations of dilated convolutions to obtain larger and more effective receptive fields.
To better train this efficient generator, except for frequently-used VGG feature matching loss, we design a novel self-guided regression loss.
We also employ a discriminator with local and global branches to ensure local-global contents consistency.
arXiv Detail & Related papers (2020-02-07T03:45:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.