Further Analysis of Outlier Detection with Deep Generative Models
- URL: http://arxiv.org/abs/2010.13064v1
- Date: Sun, 25 Oct 2020 08:20:38 GMT
- Title: Further Analysis of Outlier Detection with Deep Generative Models
- Authors: Ziyu Wang, Bin Dai, David Wipf and Jun Zhu
- Abstract summary: Deep generative models can frequently assign a higher likelihood to outliers.
We present a possible explanation for this phenomenon, starting from the observation that a model's typical set and high-density region may not conincide.
We also conduct additional experiments to help disentangle the impact of low-level texture versus high-level semantics in differentiating outliers.
- Score: 30.37180598197441
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The recent, counter-intuitive discovery that deep generative models (DGMs)
can frequently assign a higher likelihood to outliers has implications for both
outlier detection applications as well as our overall understanding of
generative modeling. In this work, we present a possible explanation for this
phenomenon, starting from the observation that a model's typical set and
high-density region may not conincide. From this vantage point we propose a
novel outlier test, the empirical success of which suggests that the failure of
existing likelihood-based outlier tests does not necessarily imply that the
corresponding generative model is uncalibrated. We also conduct additional
experiments to help disentangle the impact of low-level texture versus
high-level semantics in differentiating outliers. In aggregate, these results
suggest that modifications to the standard evaluation practices and benchmarks
commonly applied in the literature are needed.
Related papers
- Demystifying amortized causal discovery with transformers [21.058343547918053]
Supervised learning approaches for causal discovery from observational data often achieve competitive performance.
In this work, we investigate CSIvA, a transformer-based model promising to train on synthetic data and transfer to real data.
We bridge the gap with existing identifiability theory and show that constraints on the training data distribution implicitly define a prior on the test observations.
arXiv Detail & Related papers (2024-05-27T08:17:49Z) - The Surprising Harmfulness of Benign Overfitting for Adversarial
Robustness [13.120373493503772]
We prove a surprising result that even if the ground truth itself is robust to adversarial examples, the benignly overfitted model is benign in terms of the standard'' out-of-sample risk objective.
Our finding provides theoretical insights into the puzzling phenomenon observed in practice, where the true target function (e.g., human) is robust against adverasrial attack, while beginly overfitted neural networks lead to models that are not robust.
arXiv Detail & Related papers (2024-01-19T15:40:46Z) - Explainable Deep Few-shot Anomaly Detection with Deviation Networks [123.46611927225963]
We introduce a novel weakly-supervised anomaly detection framework to train detection models.
The proposed approach learns discriminative normality by leveraging the labeled anomalies and a prior probability.
Our model is substantially more sample-efficient and robust, and performs significantly better than state-of-the-art competing methods in both closed-set and open-set settings.
arXiv Detail & Related papers (2021-08-01T14:33:17Z) - Predicting Unreliable Predictions by Shattering a Neural Network [145.3823991041987]
Piecewise linear neural networks can be split into subfunctions.
Subfunctions have their own activation pattern, domain, and empirical error.
Empirical error for the full network can be written as an expectation over subfunctions.
arXiv Detail & Related papers (2021-06-15T18:34:41Z) - Robust Out-of-Distribution Detection on Deep Probabilistic Generative
Models [0.06372261626436676]
Out-of-distribution (OOD) detection is an important task in machine learning systems.
Deep probabilistic generative models facilitate OOD detection by estimating the likelihood of a data sample.
We propose a new detection metric that operates without outlier exposure.
arXiv Detail & Related papers (2021-06-15T06:36:10Z) - Perfect density models cannot guarantee anomaly detection [9.64157020816848]
In this paper, we take a closer look at the behavior of distribution densities and show that these quantities carry less meaningful information than previously thought.
We conclude that the use of these likelihoods for out-of-distribution detection relies on strong and implicit hypotheses.
arXiv Detail & Related papers (2020-12-07T15:50:11Z) - Understanding Classifier Mistakes with Generative Models [88.20470690631372]
Deep neural networks are effective on supervised learning tasks, but have been shown to be brittle.
In this paper, we leverage generative models to identify and characterize instances where classifiers fail to generalize.
Our approach is agnostic to class labels from the training set which makes it applicable to models trained in a semi-supervised way.
arXiv Detail & Related papers (2020-10-05T22:13:21Z) - Goal-directed Generation of Discrete Structures with Conditional
Generative Models [85.51463588099556]
We introduce a novel approach to directly optimize a reinforcement learning objective, maximizing an expected reward.
We test our methodology on two tasks: generating molecules with user-defined properties and identifying short python expressions which evaluate to a given target value.
arXiv Detail & Related papers (2020-10-05T20:03:13Z) - Good Classifiers are Abundant in the Interpolating Regime [64.72044662855612]
We develop a methodology to compute precisely the full distribution of test errors among interpolating classifiers.
We find that test errors tend to concentrate around a small typical value $varepsilon*$, which deviates substantially from the test error of worst-case interpolating model.
Our results show that the usual style of analysis in statistical learning theory may not be fine-grained enough to capture the good generalization performance observed in practice.
arXiv Detail & Related papers (2020-06-22T21:12:31Z) - Goodness-of-Fit Test for Mismatched Self-Exciting Processes [18.892845399295254]
We develop a GOF test for generative models of self-exciting processes by making a new connection to this problem with the classical statistical theory of Quasi-likelihood estimator (QMLE)
We present a non-parametric self-normalizing statistic for the GOF test: the Generalized Score (GS) statistics, and explicitly capture the model misspecification when establishing the distribution of the GS statistic.
arXiv Detail & Related papers (2020-06-16T18:31:33Z) - Unsupervised Anomaly Detection with Adversarial Mirrored AutoEncoders [51.691585766702744]
We propose a variant of Adversarial Autoencoder which uses a mirrored Wasserstein loss in the discriminator to enforce better semantic-level reconstruction.
We put forward an alternative measure of anomaly score to replace the reconstruction-based metric.
Our method outperforms the current state-of-the-art methods for anomaly detection on several OOD detection benchmarks.
arXiv Detail & Related papers (2020-03-24T08:26:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.