Exploring the Optimization Objective of One-Class Classification for
Anomaly Detection
- URL: http://arxiv.org/abs/2308.11898v2
- Date: Fri, 25 Aug 2023 06:38:23 GMT
- Title: Exploring the Optimization Objective of One-Class Classification for
Anomaly Detection
- Authors: Han Gao, Huiyuan Luo, Fei Shen, Zhengtao Zhang
- Abstract summary: One-class classification (OCC) is a longstanding method for anomaly detection.
In this work, we conduct a thorough investigation into the optimization objective of OCC.
We unveil a key insights: any space with the suitable norm can serve as an equivalent substitute for the hypersphere center.
This novel insight sparks a simple and data-agnostic deep one-class classification method.
- Score: 2.9266769103356305
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: One-class classification (OCC) is a longstanding method for anomaly
detection. With the powerful representation capability of the pre-trained
backbone, OCC methods have witnessed significant performance improvements.
Typically, most of these OCC methods employ transfer learning to enhance the
discriminative nature of the pre-trained backbone's features, thus achieving
remarkable efficacy. While most current approaches emphasize feature transfer
strategies, we argue that the optimization objective space within OCC methods
could also be an underlying critical factor influencing performance. In this
work, we conducted a thorough investigation into the optimization objective of
OCC. Through rigorous theoretical analysis and derivation, we unveil a key
insights: any space with the suitable norm can serve as an equivalent
substitute for the hypersphere center, without relying on the distribution
assumption of training samples. Further, we provide guidelines for determining
the feasible domain of norms for the OCC optimization objective. This novel
insight sparks a simple and data-agnostic deep one-class classification method.
Our method is straightforward, with a single 1x1 convolutional layer as a
trainable projector and any space with suitable norm as the optimization
objective. Extensive experiments validate the reliability and efficacy of our
findings and the corresponding methodology, resulting in state-of-the-art
performance in both one-class classification and industrial vision anomaly
detection and segmentation tasks.
Related papers
- Enhancing Robustness of Vision-Language Models through Orthogonality Learning and Self-Regularization [77.62516752323207]
We introduce an orthogonal fine-tuning method for efficiently fine-tuning pretrained weights and enabling enhanced robustness and generalization.
A self-regularization strategy is further exploited to maintain the stability in terms of zero-shot generalization of VLMs, dubbed OrthSR.
For the first time, we revisit the CLIP and CoOp with our method to effectively improve the model on few-shot image classficiation scenario.
arXiv Detail & Related papers (2024-07-11T10:35:53Z) - Confidence-aware Contrastive Learning for Selective Classification [20.573658672018066]
This work provides a generalization bound for selective classification, disclosing that optimizing feature layers helps improve the performance of selective classification.
Inspired by this theory, we propose to explicitly improve the selective classification model at the feature level for the first time, leading to a novel Confidence-aware Contrastive Learning method for Selective Classification, CCL-SC.
arXiv Detail & Related papers (2024-06-07T08:43:53Z) - GCC: Generative Calibration Clustering [55.44944397168619]
We propose a novel Generative Clustering (GCC) method to incorporate feature learning and augmentation into clustering procedure.
First, we develop a discrimirative feature alignment mechanism to discover intrinsic relationship across real and generated samples.
Second, we design a self-supervised metric learning to generate more reliable cluster assignment.
arXiv Detail & Related papers (2024-04-14T01:51:11Z) - From Function to Distribution Modeling: A PAC-Generative Approach to
Offline Optimization [30.689032197123755]
This paper considers the problem of offline optimization, where the objective function is unknown except for a collection of offline" data examples.
Instead of learning and then optimizing the unknown objective function, we take on a less intuitive but more direct view that optimization can be thought of as a process of sampling from a generative model.
arXiv Detail & Related papers (2024-01-04T01:32:50Z) - Deterministic Langevin Unconstrained Optimization with Normalizing Flows [3.988614978933934]
We introduce a global, free surrogate optimization strategy for black-box functions inspired by the Fokker-Planck and Langevin equations.
We demonstrate superior competitive progress toward objective optima on standard synthetic test functions.
arXiv Detail & Related papers (2023-10-01T17:46:20Z) - Taxonomy Adaptive Cross-Domain Adaptation in Medical Imaging via
Optimization Trajectory Distillation [73.83178465971552]
The success of automated medical image analysis depends on large-scale and expert-annotated training sets.
Unsupervised domain adaptation (UDA) has been raised as a promising approach to alleviate the burden of labeled data collection.
We propose optimization trajectory distillation, a unified approach to address the two technical challenges from a new perspective.
arXiv Detail & Related papers (2023-07-27T08:58:05Z) - Self-training through Classifier Disagreement for Cross-Domain Opinion
Target Extraction [62.41511766918932]
Opinion target extraction (OTE) or aspect extraction (AE) is a fundamental task in opinion mining.
Recent work focus on cross-domain OTE, which is typically encountered in real-world scenarios.
We propose a new SSL approach that opts for selecting target samples whose model output from a domain-specific teacher and student network disagrees on the unlabelled target data.
arXiv Detail & Related papers (2023-02-28T16:31:17Z) - Domain Adaptation with Adversarial Training on Penultimate Activations [82.9977759320565]
Enhancing model prediction confidence on unlabeled target data is an important objective in Unsupervised Domain Adaptation (UDA)
We show that this strategy is more efficient and better correlated with the objective of boosting prediction confidence than adversarial training on input images or intermediate features.
arXiv Detail & Related papers (2022-08-26T19:50:46Z) - Boosting the Generalization Capability in Cross-Domain Few-shot Learning
via Noise-enhanced Supervised Autoencoder [23.860842627883187]
We teach the model to capture broader variations of the feature distributions with a novel noise-enhanced supervised autoencoder (NSAE)
NSAE trains the model by jointly reconstructing inputs and predicting the labels of inputs as well as their reconstructed pairs.
We also take advantage of NSAE structure and propose a two-step fine-tuning procedure that achieves better adaption and improves classification performance in the target domain.
arXiv Detail & Related papers (2021-08-11T04:45:56Z) - MCDAL: Maximum Classifier Discrepancy for Active Learning [74.73133545019877]
Recent state-of-the-art active learning methods have mostly leveraged Generative Adversarial Networks (GAN) for sample acquisition.
We propose in this paper a novel active learning framework that we call Maximum Discrepancy for Active Learning (MCDAL)
In particular, we utilize two auxiliary classification layers that learn tighter decision boundaries by maximizing the discrepancies among them.
arXiv Detail & Related papers (2021-07-23T06:57:08Z) - PANDA: Adapting Pretrained Features for Anomaly Detection and
Segmentation [34.98371632913735]
We show that combining pretrained features with simple anomaly detection and segmentation methods convincingly outperforms state-of-the-art methods.
In order to obtain further performance gains, we adapt pretrained features to the target distribution.
arXiv Detail & Related papers (2020-10-12T17:52:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.