Statistical Evaluation of Anomaly Detectors for Sequences
- URL: http://arxiv.org/abs/2008.05788v1
- Date: Thu, 13 Aug 2020 10:07:27 GMT
- Title: Statistical Evaluation of Anomaly Detectors for Sequences
- Authors: Erik Scharw\"achter and Emmanuel M\"uller
- Abstract summary: We formalize a notion of precision and recall with temporal tolerance for point-based anomaly detection in sequential data.
We show how to obtain null distributions for the two measures to assess the statistical significance of reported results.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Although precision and recall are standard performance measures for anomaly
detection, their statistical properties in sequential detection settings are
poorly understood. In this work, we formalize a notion of precision and recall
with temporal tolerance for point-based anomaly detection in sequential data.
These measures are based on time-tolerant confusion matrices that may be used
to compute time-tolerant variants of many other standard measures. However,
care has to be taken to preserve interpretability. We perform a statistical
simulation study to demonstrate that precision and recall may overestimate the
performance of a detector, when computed with temporal tolerance. To alleviate
this problem, we show how to obtain null distributions for the two measures to
assess the statistical significance of reported results.
Related papers
- An Evaluation of Real-time Adaptive Sampling Change Point Detection Algorithm using KCUSUM [4.610597418629838]
We introduce the Kernel-based Cumulative Sum (KCUSUM) algorithm, a non-parametric extension of the traditional Cumulative Sum (CUSUM) method.
KCUSUM splits itself by comparing incoming samples directly with reference samples and computes a statistic grounded in the Maximum Mean Discrepancy (MMD) non-parametric framework.
We discuss real-world use cases from scientific simulations such as NWChem CODAR and protein folding data, demonstrating KCUSUM's practical effectiveness in online change point detection.
arXiv Detail & Related papers (2024-02-15T19:45:24Z) - Score Matching-based Pseudolikelihood Estimation of Neural Marked
Spatio-Temporal Point Process with Uncertainty Quantification [59.81904428056924]
We introduce SMASH: a Score MAtching estimator for learning markedPs with uncertainty quantification.
Specifically, our framework adopts a normalization-free objective by estimating the pseudolikelihood of markedPs through score-matching.
The superior performance of our proposed framework is demonstrated through extensive experiments in both event prediction and uncertainty quantification.
arXiv Detail & Related papers (2023-10-25T02:37:51Z) - Adaptive Thresholding Heuristic for KPI Anomaly Detection [1.57731592348751]
A plethora of outlier detectors have been explored in the time series domain, however, in a business sense, not all outliers are anomalies of interest.
This article proposes an Adaptive Thresholding Heuristic (ATH) to dynamically adjust the detection threshold based on the local properties of the data distribution and adapt to changes in time series patterns.
Experimental results show that ATH is efficient making it scalable for near real time anomaly detection and flexible with forecasters and outlier detectors.
arXiv Detail & Related papers (2023-08-21T06:45:28Z) - SoftED: Metrics for Soft Evaluation of Time Series Event Detection [4.263111781491367]
Time series event detection methods are evaluated mainly by standard classification metrics that focus solely on detection accuracy.
Inaccuracy in detecting an event can often result from its preceding or delayed effects reflected in neighboring detections.
This paper introduces SoftED metrics, a new set of metrics designed for soft evaluating event detection methods.
arXiv Detail & Related papers (2023-04-02T03:27:31Z) - The Implicit Delta Method [61.36121543728134]
In this paper, we propose an alternative, the implicit delta method, which works by infinitesimally regularizing the training loss of uncertainty.
We show that the change in the evaluation due to regularization is consistent for the variance of the evaluation estimator, even when the infinitesimal change is approximated by a finite difference.
arXiv Detail & Related papers (2022-11-11T19:34:17Z) - Local Evaluation of Time Series Anomaly Detection Algorithms [9.717823994163277]
We show that an adversary algorithm can reach high precision and recall on almost any dataset under weak assumption.
We propose a theoretically grounded, robust, parameter-free and interpretable extension to precision/recall metrics.
arXiv Detail & Related papers (2022-06-27T10:18:41Z) - TACTiS: Transformer-Attentional Copulas for Time Series [76.71406465526454]
estimation of time-varying quantities is a fundamental component of decision making in fields such as healthcare and finance.
We propose a versatile method that estimates joint distributions using an attention-based decoder.
We show that our model produces state-of-the-art predictions on several real-world datasets.
arXiv Detail & Related papers (2022-02-07T21:37:29Z) - A Statistical Analysis of Summarization Evaluation Metrics using
Resampling Methods [60.04142561088524]
We find that the confidence intervals are rather wide, demonstrating high uncertainty in how reliable automatic metrics truly are.
Although many metrics fail to show statistical improvements over ROUGE, two recent works, QAEval and BERTScore, do in some evaluation settings.
arXiv Detail & Related papers (2021-03-31T18:28:14Z) - Performance metrics for intervention-triggering prediction models do not
reflect an expected reduction in outcomes from using the model [71.9860741092209]
Clinical researchers often select among and evaluate risk prediction models.
Standard metrics calculated from retrospective data are only related to model utility under certain assumptions.
When predictions are delivered repeatedly throughout time, the relationship between standard metrics and utility is further complicated.
arXiv Detail & Related papers (2020-06-02T16:26:49Z) - Learning to Predict Error for MRI Reconstruction [67.76632988696943]
We demonstrate that predictive uncertainty estimated by the current methods does not highly correlate with prediction error.
We propose a novel method that estimates the target labels and magnitude of the prediction error in two steps.
arXiv Detail & Related papers (2020-02-13T15:55:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.