The Clever Hans Effect in Anomaly Detection
- URL: http://arxiv.org/abs/2006.10609v1
- Date: Thu, 18 Jun 2020 15:27:05 GMT
- Title: The Clever Hans Effect in Anomaly Detection
- Authors: Jacob Kauffmann, Lukas Ruff, Gr\'egoire Montavon, Klaus-Robert
M\"uller
- Abstract summary: The 'Clever Hans' effect occurs when the learned model produces correct predictions based on the 'wrong' features.
This paper will contribute an explainable AI (XAI) procedure that can highlight the relevant features used by popular anomaly detection models.
- Score: 3.278983768346415
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The 'Clever Hans' effect occurs when the learned model produces correct
predictions based on the 'wrong' features. This effect which undermines the
generalization capability of an ML model and goes undetected by standard
validation techniques has been frequently observed for supervised learning
where the training algorithm leverages spurious correlations in the data. The
question whether Clever Hans also occurs in unsupervised learning, and in which
form, has received so far almost no attention. Therefore, this paper will
contribute an explainable AI (XAI) procedure that can highlight the relevant
features used by popular anomaly detection models of different type. Our
analysis reveals that the Clever Hans effect is widespread in anomaly detection
and occurs in many (unexpected) forms. Interestingly, the observed Clever Hans
effects are in this case not so much due to the data, but due to the anomaly
detection models themselves whose structure makes them unable to detect the
truly relevant features, even though vast amounts of data points are available.
Overall, our work contributes a warning against an unrestrained use of existing
anomaly detection models in practical applications, but it also points at a
possible way out of the Clever Hans dilemma, specifically, by allowing multiple
anomaly models to mutually cancel their individual structural weaknesses to
jointly produce a better and more trustworthy anomaly detector.
Related papers
- Can I trust my anomaly detection system? A case study based on explainable AI [0.4416503115535552]
This case study explores the robustness of an anomaly detection system based on variational autoencoder generative models.
The goal is to get a different perspective on the real performances of anomaly detectors that use reconstruction differences.
arXiv Detail & Related papers (2024-07-29T12:39:07Z) - Multiple Descents in Unsupervised Learning: The Role of Noise, Domain Shift and Anomalies [14.399035468023161]
We study the presence of double descent in unsupervised learning, an area that has received little attention and is not yet fully understood.
We use synthetic and real data and identify model-wise, epoch-wise, and sample-wise double descent for various applications.
arXiv Detail & Related papers (2024-06-17T16:24:23Z) - Towards a Unified Framework of Clustering-based Anomaly Detection [18.30208347233284]
Unsupervised Anomaly Detection (UAD) plays a crucial role in identifying abnormal patterns within data without labeled examples.
We propose a novel probabilistic mixture model for anomaly detection to establish a theoretical connection among representation learning, clustering, and anomaly detection.
We have devised an improved anomaly score that more effectively harnesses the combined power of representation learning and clustering.
arXiv Detail & Related papers (2024-06-01T14:30:12Z) - Prototypes as Explanation for Time Series Anomaly Detection [6.051581987453758]
This paper proposes ProtoAD, using prototypes as the example-based explanation for the state of regular patterns during anomaly detection.
By visualizing both the latent space and input space prototypes, we intuitively demonstrate how regular data are modeled and why specific patterns are considered abnormal.
arXiv Detail & Related papers (2023-07-04T09:40:30Z) - Causality-Based Multivariate Time Series Anomaly Detection [63.799474860969156]
We formulate the anomaly detection problem from a causal perspective and view anomalies as instances that do not follow the regular causal mechanism to generate the multivariate data.
We then propose a causality-based anomaly detection approach, which first learns the causal structure from data and then infers whether an instance is an anomaly relative to the local causal mechanism.
We evaluate our approach with both simulated and public datasets as well as a case study on real-world AIOps applications.
arXiv Detail & Related papers (2022-06-30T06:00:13Z) - Catching Both Gray and Black Swans: Open-set Supervised Anomaly
Detection [90.32910087103744]
A few labeled anomaly examples are often available in many real-world applications.
These anomaly examples provide valuable knowledge about the application-specific abnormality.
Those anomalies seen during training often do not illustrate every possible class of anomaly.
This paper tackles open-set supervised anomaly detection.
arXiv Detail & Related papers (2022-03-28T05:21:37Z) - SLA$^2$P: Self-supervised Anomaly Detection with Adversarial
Perturbation [77.71161225100927]
Anomaly detection is a fundamental yet challenging problem in machine learning.
We propose a novel and powerful framework, dubbed as SLA$2$P, for unsupervised anomaly detection.
arXiv Detail & Related papers (2021-11-25T03:53:43Z) - Explainable Deep Few-shot Anomaly Detection with Deviation Networks [123.46611927225963]
We introduce a novel weakly-supervised anomaly detection framework to train detection models.
The proposed approach learns discriminative normality by leveraging the labeled anomalies and a prior probability.
Our model is substantially more sample-efficient and robust, and performs significantly better than state-of-the-art competing methods in both closed-set and open-set settings.
arXiv Detail & Related papers (2021-08-01T14:33:17Z) - Understanding the Effect of Bias in Deep Anomaly Detection [15.83398707988473]
Anomaly detection presents a unique challenge in machine learning, due to the scarcity of labeled anomaly data.
Recent work attempts to mitigate such problems by augmenting training of deep anomaly detection models with additional labeled anomaly samples.
In this paper, we aim to understand the effect of a biased anomaly set on anomaly detection.
arXiv Detail & Related papers (2021-05-16T03:55:02Z) - Learning Causal Models Online [103.87959747047158]
Predictive models can rely on spurious correlations in the data for making predictions.
One solution for achieving strong generalization is to incorporate causal structures in the models.
We propose an online algorithm that continually detects and removes spurious features.
arXiv Detail & Related papers (2020-06-12T20:49:20Z) - Deep Weakly-supervised Anomaly Detection [118.55172352231381]
Pairwise Relation prediction Network (PReNet) learns pairwise relation features and anomaly scores.
PReNet can detect any seen/unseen abnormalities that fit the learned pairwise abnormal patterns.
Empirical results on 12 real-world datasets show that PReNet significantly outperforms nine competing methods in detecting seen and unseen anomalies.
arXiv Detail & Related papers (2019-10-30T00:40:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.