MuSc: Zero-Shot Industrial Anomaly Classification and Segmentation with
Mutual Scoring of the Unlabeled Images
- URL: http://arxiv.org/abs/2401.16753v1
- Date: Tue, 30 Jan 2024 05:16:52 GMT
- Title: MuSc: Zero-Shot Industrial Anomaly Classification and Segmentation with
Mutual Scoring of the Unlabeled Images
- Authors: Xurui Li, Ziming Huang, Feng Xue, Yu Zhou
- Abstract summary: We study zero-shot anomaly classification (AC) and segmentation (AS) in industrial vision.
We leverage a discriminative characteristic to design a novel zero-shot AC/AS method by Mutual Scoring (MuSc) of the unlabeled images.
We present an optimization approach named Re-scoring with Constrained Image-level Neighborhood (RsCIN) for image-level anomaly classification.
- Score: 12.48347948647802
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper studies zero-shot anomaly classification (AC) and segmentation
(AS) in industrial vision. We reveal that the abundant normal and abnormal cues
implicit in unlabeled test images can be exploited for anomaly determination,
which is ignored by prior methods. Our key observation is that for the
industrial product images, the normal image patches could find a relatively
large number of similar patches in other unlabeled images, while the abnormal
ones only have a few similar patches. We leverage such a discriminative
characteristic to design a novel zero-shot AC/AS method by Mutual Scoring
(MuSc) of the unlabeled images, which does not need any training or prompts.
Specifically, we perform Local Neighborhood Aggregation with Multiple Degrees
(LNAMD) to obtain the patch features that are capable of representing anomalies
in varying sizes. Then we propose the Mutual Scoring Mechanism (MSM) to
leverage the unlabeled test images to assign the anomaly score to each other.
Furthermore, we present an optimization approach named Re-scoring with
Constrained Image-level Neighborhood (RsCIN) for image-level anomaly
classification to suppress the false positives caused by noises in normal
images. The superior performance on the challenging MVTec AD and VisA datasets
demonstrates the effectiveness of our approach. Compared with the
state-of-the-art zero-shot approaches, MuSc achieves a $\textbf{21.1%}$ PRO
absolute gain (from 72.7% to 93.8%) on MVTec AD, a $\textbf{19.4%}$ pixel-AP
gain and a $\textbf{14.7%}$ pixel-AUROC gain on VisA. In addition, our
zero-shot approach outperforms most of the few-shot approaches and is
comparable to some one-class methods. Code is available at
https://github.com/xrli-U/MuSc.
Related papers
- FADE: Few-shot/zero-shot Anomaly Detection Engine using Large Vision-Language Model [0.9226774742769024]
Few-shot/zero-shot anomaly detection is important for quality inspection in the manufacturing industry.
We propose the Few-shot/zero-shot Anomaly Engine Detection (FADE) which leverages the vision-language CLIP model and adjusts it for the purpose of anomaly detection.
FADE outperforms other state-of-the-art methods in anomaly segmentation with pixel-AUROC of 89.6% (91.5%) in zero-shot and 95.4% (97.5%) in 1-normal-shot.
arXiv Detail & Related papers (2024-08-31T23:05:56Z) - Few-Shot Anomaly Detection via Category-Agnostic Registration Learning [65.64252994254268]
Most existing anomaly detection methods require a dedicated model for each category.
This article proposes a novel few-shot AD (FSAD) framework.
It is the first FSAD method that requires no model fine-tuning for novel categories.
arXiv Detail & Related papers (2024-06-13T05:01:13Z) - Learning to Rank Patches for Unbiased Image Redundancy Reduction [80.93989115541966]
Images suffer from heavy spatial redundancy because pixels in neighboring regions are spatially correlated.
Existing approaches strive to overcome this limitation by reducing less meaningful image regions.
We propose a self-supervised framework for image redundancy reduction called Learning to Rank Patches.
arXiv Detail & Related papers (2024-03-31T13:12:41Z) - Rethinking Image Forgery Detection via Contrastive Learning and
Unsupervised Clustering [26.923409536155166]
We propose FOrensic ContrAstive cLustering (FOCAL) method for image forgery detection.
FOCAL is based on contrastive learning and unsupervised clustering.
Results show FOCAL significantly outperforms state-of-the-art competing algorithms.
arXiv Detail & Related papers (2023-08-18T05:05:30Z) - AMAE: Adaptation of Pre-Trained Masked Autoencoder for Dual-Distribution
Anomaly Detection in Chest X-Rays [17.91123470181453]
We propose AMAE, a two-stage algorithm for adaptation of the pre-trained masked autoencoder (MAE)
AMAE leads to consistent performance gains over competing self-supervised and dual distribution anomaly detection methods.
arXiv Detail & Related papers (2023-07-24T12:03:50Z) - Optimal transport meets noisy label robust loss and MixUp regularization
for domain adaptation [13.080485957000462]
Deep neural networks trained on a source training set perform poorly on target images which do not belong to the training domain.
One strategy to improve these performances is to align the source and target image distributions in an embedded space using optimal transport (OT)
We propose to couple the MixUp regularization citepzhang 2018mixup with a loss that is robust to noisy labels in order to improve domain adaptation performance.
arXiv Detail & Related papers (2022-06-22T15:40:52Z) - Exposing Outlier Exposure: What Can Be Learned From Few, One, and Zero
Outlier Images [26.283734474660484]
We show that specialized AD learning methods seem actually superfluous and huge corpora of data expendable.
We investigate this phenomenon and reveal that one-class methods are more robust towards the particular choice of training outliers.
arXiv Detail & Related papers (2022-05-23T17:23:15Z) - mc-BEiT: Multi-choice Discretization for Image BERT Pre-training [52.04866462439979]
Image BERT pre-training with masked image modeling (MIM) is a popular practice to cope with self-supervised representation learning.
We introduce an improved BERT-style image pre-training method, namely mc-BEiT, which performs MIM proxy tasks towards eased and refined multi-choice training objectives.
arXiv Detail & Related papers (2022-03-29T09:08:18Z) - Mixed Supervision Learning for Whole Slide Image Classification [88.31842052998319]
We propose a mixed supervision learning framework for super high-resolution images.
During the patch training stage, this framework can make use of coarse image-level labels to refine self-supervised learning.
A comprehensive strategy is proposed to suppress pixel-level false positives and false negatives.
arXiv Detail & Related papers (2021-07-02T09:46:06Z) - A Hierarchical Transformation-Discriminating Generative Model for Few
Shot Anomaly Detection [93.38607559281601]
We devise a hierarchical generative model that captures the multi-scale patch distribution of each training image.
The anomaly score is obtained by aggregating the patch-based votes of the correct transformation across scales and image regions.
arXiv Detail & Related papers (2021-04-29T17:49:48Z) - Permuted AdaIN: Reducing the Bias Towards Global Statistics in Image
Classification [97.81205777897043]
Recent work has shown that convolutional neural network classifiers overly rely on texture at the expense of shape cues.
We make a similar but different distinction between shape and local image cues, on the one hand, and global image statistics, on the other.
Our method, called Permuted Adaptive Instance Normalization (pAdaIN), reduces the representation of global statistics in the hidden layers of image classifiers.
arXiv Detail & Related papers (2020-10-09T16:38:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.