Shaken, and Stirred: Long-Range Dependencies Enable Robust Outlier
Detection with PixelCNN++
- URL: http://arxiv.org/abs/2208.13579v2
- Date: Sat, 20 May 2023 20:14:06 GMT
- Title: Shaken, and Stirred: Long-Range Dependencies Enable Robust Outlier
Detection with PixelCNN++
- Authors: Barath Mohan Umapathi, Kushal Chauhan, Pradeep Shenoy, Devarajan
Sridharan
- Abstract summary: We show that biases in PixelCNN++ likelihoods arise primarily from predictions based on local dependencies.
We propose two families of transformations -- stirring'' and shaking'' -- which ameliorate low-level biases and isolate the contribution of long-range dependencies.
We test our approaches extensively with five grayscale and six natural image datasets and show that they achieve or exceed state-of-the-art outlier detection.
- Score: 6.736754991468853
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Reliable outlier detection is critical for real-world deployment of deep
learning models. Although extensively studied, likelihoods produced by deep
generative models have been largely dismissed as being impractical for outlier
detection. First, deep generative model likelihoods are readily biased by
low-level input statistics. Second, many recent solutions for correcting these
biases are computationally expensive, or do not generalize well to complex,
natural datasets. Here, we explore outlier detection with a state-of-the-art
deep autoregressive model: PixelCNN++. We show that biases in PixelCNN++
likelihoods arise primarily from predictions based on local dependencies. We
propose two families of bijective transformations -- ``stirring'' and
``shaking'' -- which ameliorate low-level biases and isolate the contribution
of long-range dependencies to PixelCNN++ likelihoods. These transformations are
inexpensive and readily computed at evaluation time. We test our approaches
extensively with five grayscale and six natural image datasets and show that
they achieve or exceed state-of-the-art outlier detection, particularly on
datasets with complex, natural images. We also show that our solutions work
well with other types of generative models (generative flows and variational
autoencoders) and that their efficacy is governed by each model's reliance on
local dependencies. In sum, lightweight remedies suffice to achieve robust
outlier detection on image data with deep generative models.
Related papers
- Open-Set Deepfake Detection: A Parameter-Efficient Adaptation Method with Forgery Style Mixture [58.60915132222421]
We introduce an approach that is both general and parameter-efficient for face forgery detection.
We design a forgery-style mixture formulation that augments the diversity of forgery source domains.
We show that the designed model achieves state-of-the-art generalizability with significantly reduced trainable parameters.
arXiv Detail & Related papers (2024-08-23T01:53:36Z) - Exploring the Effectiveness of Dataset Synthesis: An application of
Apple Detection in Orchards [68.95806641664713]
We explore the usability of Stable Diffusion 2.1-base for generating synthetic datasets of apple trees for object detection.
We train a YOLOv5m object detection model to predict apples in a real-world apple detection dataset.
Results demonstrate that the model trained on generated data is slightly underperforming compared to a baseline model trained on real-world images.
arXiv Detail & Related papers (2023-06-20T09:46:01Z) - Semi-supervised Deep Learning for Image Classification with Distribution
Mismatch: A Survey [1.5469452301122175]
Deep learning models rely on the abundance of labelled observations to train a prospective model.
It is expensive to gather labelled observations of data, making the usage of deep learning models not ideal.
In many situations different unlabelled data sources might be available.
This raises the risk of a significant distribution mismatch between the labelled and unlabelled datasets.
arXiv Detail & Related papers (2022-03-01T02:46:00Z) - X-model: Improving Data Efficiency in Deep Learning with A Minimax Model [78.55482897452417]
We aim at improving data efficiency for both classification and regression setups in deep learning.
To take the power of both worlds, we propose a novel X-model.
X-model plays a minimax game between the feature extractor and task-specific heads.
arXiv Detail & Related papers (2021-10-09T13:56:48Z) - Efficient remedies for outlier detection with variational autoencoders [8.80692072928023]
Likelihoods computed by deep generative models are a candidate metric for outlier detection with unlabeled data.
We show that a theoretically-grounded correction readily ameliorates a key bias with VAE likelihood estimates.
We also show that the variance of the likelihoods computed over an ensemble of VAEs also enables robust outlier detection.
arXiv Detail & Related papers (2021-08-19T16:00:58Z) - Robust Out-of-Distribution Detection on Deep Probabilistic Generative
Models [0.06372261626436676]
Out-of-distribution (OOD) detection is an important task in machine learning systems.
Deep probabilistic generative models facilitate OOD detection by estimating the likelihood of a data sample.
We propose a new detection metric that operates without outlier exposure.
arXiv Detail & Related papers (2021-06-15T06:36:10Z) - Contemplating real-world object classification [53.10151901863263]
We reanalyze the ObjectNet dataset recently proposed by Barbu et al. containing objects in daily life situations.
We find that applying deep models to the isolated objects, rather than the entire scene as is done in the original paper, results in around 20-30% performance improvement.
arXiv Detail & Related papers (2021-03-08T23:29:59Z) - Goal-directed Generation of Discrete Structures with Conditional
Generative Models [85.51463588099556]
We introduce a novel approach to directly optimize a reinforcement learning objective, maximizing an expected reward.
We test our methodology on two tasks: generating molecules with user-defined properties and identifying short python expressions which evaluate to a given target value.
arXiv Detail & Related papers (2020-10-05T20:03:13Z) - Understanding Anomaly Detection with Deep Invertible Networks through
Hierarchies of Distributions and Features [4.25227087152716]
Convolutional networks learn similar low-level feature distributions when trained on any natural image dataset.
When the discriminative features between inliers and outliers are on a high-level, anomaly detection becomes particularly challenging.
We propose two methods to remove the negative impact of model bias and domain prior on detecting high-level differences.
arXiv Detail & Related papers (2020-06-18T20:56:14Z) - Diversity inducing Information Bottleneck in Model Ensembles [73.80615604822435]
In this paper, we target the problem of generating effective ensembles of neural networks by encouraging diversity in prediction.
We explicitly optimize a diversity inducing adversarial loss for learning latent variables and thereby obtain diversity in the output predictions necessary for modeling multi-modal data.
Compared to the most competitive baselines, we show significant improvements in classification accuracy, under a shift in the data distribution.
arXiv Detail & Related papers (2020-03-10T03:10:41Z) - Granular Learning with Deep Generative Models using Highly Contaminated
Data [0.0]
An approach to utilize recent advances in deep generative models for anomaly detection in a granular sense on a real-world image dataset with quality issues is detailed.
The approach is completely unsupervised (no annotations available) but qualitatively shown to provide accurate semantic labeling for images.
arXiv Detail & Related papers (2020-01-06T23:22:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.