UADB: Unsupervised Anomaly Detection Booster
- URL: http://arxiv.org/abs/2306.01997v2
- Date: Tue, 26 Dec 2023 15:34:11 GMT
- Title: UADB: Unsupervised Anomaly Detection Booster
- Authors: Hangting Ye, Zhining Liu, Xinyi Shen, Wei Cao, Shun Zheng, Xiaofan
Gui, Huishuai Zhang, Yi Chang, Jiang Bian
- Abstract summary: Unsupervised Anomaly Detection (UAD) is a key data mining problem owing to its wide real-world applications.
No single assumption can describe such complexity and be valid in all scenarios.
We propose a general UAD Booster (UADB) that empowers any UAD models with adaptability to different data.
- Score: 29.831918685340433
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Unsupervised Anomaly Detection (UAD) is a key data mining problem owing to
its wide real-world applications. Due to the complete absence of supervision
signals, UAD methods rely on implicit assumptions about anomalous patterns
(e.g., scattered/sparsely/densely clustered) to detect anomalies. However,
real-world data are complex and vary significantly across different domains. No
single assumption can describe such complexity and be valid in all scenarios.
This is also confirmed by recent research that shows no UAD method is
omnipotent. Based on above observations, instead of searching for a magic
universal winner assumption, we seek to design a general UAD Booster (UADB)
that empowers any UAD models with adaptability to different data. This is a
challenging task given the heterogeneous model structures and assumptions
adopted by existing UAD methods. To achieve this, we dive deep into the UAD
problem and find that compared to normal data, anomalies (i) lack clear
structure/pattern in feature space, thus (ii) harder to learn by model without
a suitable assumption, and finally, leads to (iii) high variance between
different learners. In light of these findings, we propose to (i) distill the
knowledge of the source UAD model to an imitation learner (booster) that holds
no data assumption, then (ii) exploit the variance between them to perform
automatic correction, and thus (iii) improve the booster over the original UAD
model. We use a neural network as the booster for its strong expressive power
as a universal approximator and ability to perform flexible post-hoc tuning.
Note that UADB is a model-agnostic framework that can enhance heterogeneous UAD
models in a unified way. Extensive experiments on over 80 tabular datasets
demonstrate the effectiveness of UADB.
Related papers
- DACAD: Domain Adaptation Contrastive Learning for Anomaly Detection in Multivariate Time Series [25.434379659643707]
In time series anomaly detection, the scarcity of labeled data poses a challenge to the development of accurate models.
We propose a novel Domain Contrastive learning model for Anomaly Detection in time series (DACAD)
Our model employs supervised contrastive loss for the source domain and self-supervised contrastive triplet loss for the target domain.
arXiv Detail & Related papers (2024-04-17T11:20:14Z) - Learning Feature Inversion for Multi-class Anomaly Detection under General-purpose COCO-AD Benchmark [101.23684938489413]
Anomaly detection (AD) is often focused on detecting anomalies for industrial quality inspection and medical lesion examination.
This work first constructs a large-scale and general-purpose COCO-AD dataset by extending COCO to the AD field.
Inspired by the metrics in the segmentation field, we propose several more practical threshold-dependent AD-specific metrics.
arXiv Detail & Related papers (2024-04-16T17:38:26Z) - Self-supervised Feature Adaptation for 3D Industrial Anomaly Detection [59.41026558455904]
We focus on multi-modal anomaly detection. Specifically, we investigate early multi-modal approaches that attempted to utilize models pre-trained on large-scale visual datasets.
We propose a Local-to-global Self-supervised Feature Adaptation (LSFA) method to finetune the adaptors and learn task-oriented representation toward anomaly detection.
arXiv Detail & Related papers (2024-01-06T07:30:41Z) - Unraveling the "Anomaly" in Time Series Anomaly Detection: A
Self-supervised Tri-domain Solution [89.16750999704969]
Anomaly labels hinder traditional supervised models in time series anomaly detection.
Various SOTA deep learning techniques, such as self-supervised learning, have been introduced to tackle this issue.
We propose a novel self-supervised learning based Tri-domain Anomaly Detector (TriAD)
arXiv Detail & Related papers (2023-11-19T05:37:18Z) - Enhancing Interpretability and Generalizability in Extended Isolation Forests [5.139809663513828]
Extended Isolation Forest Feature Importance (ExIFFI) is a method that explains predictions made by Extended Isolation Forest (EIF) models.
EIF+ is designed to enhance the model's ability to detect unseen anomalies through a revised splitting strategy.
ExIFFI outperforms other unsupervised interpretability methods on 8 of 11 real-world datasets.
arXiv Detail & Related papers (2023-10-09T07:24:04Z) - DTAAD: Dual Tcn-Attention Networks for Anomaly Detection in Multivariate Time Series Data [0.0]
We propose an anomaly detection and diagnosis model, DTAAD, based on Transformer and Dual Temporal Convolutional Network (TCN)
scaling methods and feedback mechanisms are introduced to improve prediction accuracy and expand correlation differences.
Our experiments on seven public datasets validate that DTAAD exceeds the majority of currently advanced baseline methods in both detection and diagnostic performance.
arXiv Detail & Related papers (2023-02-17T06:59:45Z) - Data-Efficient and Interpretable Tabular Anomaly Detection [54.15249463477813]
We propose a novel framework that adapts a white-box model class, Generalized Additive Models, to detect anomalies.
In addition, the proposed framework, DIAD, can incorporate a small amount of labeled data to further boost anomaly detection performances in semi-supervised settings.
arXiv Detail & Related papers (2022-03-03T22:02:56Z) - SUOD: Accelerating Large-Scale Unsupervised Heterogeneous Outlier
Detection [63.253850875265115]
Outlier detection (OD) is a key machine learning (ML) task for identifying abnormal objects from general samples.
We propose a modular acceleration system, called SUOD, to address it.
arXiv Detail & Related papers (2020-03-11T00:22:50Z) - Do We Really Need to Access the Source Data? Source Hypothesis Transfer
for Unsupervised Domain Adaptation [102.67010690592011]
Unsupervised adaptationUDA (UDA) aims to leverage the knowledge learned from a labeled source dataset to solve similar tasks in a new unlabeled domain.
Prior UDA methods typically require to access the source data when learning to adapt the model.
This work tackles a practical setting where only a trained source model is available and how we can effectively utilize such a model without source data to solve UDA problems.
arXiv Detail & Related papers (2020-02-20T03:13:58Z) - Towards Out-of-Distribution Detection with Divergence Guarantee in Deep
Generative Models [22.697643259435115]
Deep generative models may assign higher likelihood to out-of-distribution (OOD) data than in-distribution (ID) data.
We prove theorems to investigate the divergences in flow-based model.
We propose two group anomaly detection methods.
arXiv Detail & Related papers (2020-02-09T09:54:12Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.