Unsupervised Outlier Detection using Memory and Contrastive Learning
- URL: http://arxiv.org/abs/2107.12642v1
- Date: Tue, 27 Jul 2021 07:35:42 GMT
- Title: Unsupervised Outlier Detection using Memory and Contrastive Learning
- Authors: Ning Huyan, Dou Quan, Xiangrong Zhang, Xuefeng Liang, Jocelyn
Chanussot, Licheng Jiao
- Abstract summary: We think outlier detection can be done in the feature space by measuring the feature distance between outliers and inliers.
We propose a framework, MCOD, using a memory module and a contrastive learning module.
Our proposed MCOD achieves a considerable performance and outperforms nine state-of-the-art methods.
- Score: 53.77693158251706
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Outlier detection is one of the most important processes taken to create
good, reliable data in machine learning. The most methods of outlier detection
leverage an auxiliary reconstruction task by assuming that outliers are more
difficult to be recovered than normal samples (inliers). However, it is not
always true, especially for auto-encoder (AE) based models. They may recover
certain outliers even outliers are not in the training data, because they do
not constrain the feature learning. Instead, we think outlier detection can be
done in the feature space by measuring the feature distance between outliers
and inliers. We then propose a framework, MCOD, using a memory module and a
contrastive learning module. The memory module constrains the consistency of
features, which represent the normal data. The contrastive learning module
learns more discriminating features, which boosts the distinction between
outliers and inliers. Extensive experiments on four benchmark datasets show
that our proposed MCOD achieves a considerable performance and outperforms nine
state-of-the-art methods.
Related papers
- Regularized Contrastive Partial Multi-view Outlier Detection [76.77036536484114]
We propose a novel method named Regularized Contrastive Partial Multi-view Outlier Detection (RCPMOD)
In this framework, we utilize contrastive learning to learn view-consistent information and distinguish outliers by the degree of consistency.
Experimental results on four benchmark datasets demonstrate that our proposed approach could outperform state-of-the-art competitors.
arXiv Detail & Related papers (2024-08-02T14:34:27Z) - Robust Capped lp-Norm Support Vector Ordinal Regression [85.84718111830752]
Ordinal regression is a specialized supervised problem where the labels show an inherent order.
Support Vector Ordinal Regression, as an outstanding ordinal regression model, is widely used in many ordinal regression tasks.
We introduce a new model, Capped $ell_p$-Norm Support Vector Ordinal Regression(CSVOR), that is robust to outliers.
arXiv Detail & Related papers (2024-04-25T13:56:05Z) - Enhancing Consistency and Mitigating Bias: A Data Replay Approach for
Incremental Learning [100.7407460674153]
Deep learning systems are prone to catastrophic forgetting when learning from a sequence of tasks.
To mitigate the problem, a line of methods propose to replay the data of experienced tasks when learning new tasks.
However, it is not expected in practice considering the memory constraint or data privacy issue.
As a replacement, data-free data replay methods are proposed by inverting samples from the classification model.
arXiv Detail & Related papers (2024-01-12T12:51:12Z) - Unsupervised Continual Anomaly Detection with Contrastively-learned
Prompt [80.43623986759691]
We introduce a novel Unsupervised Continual Anomaly Detection framework called UCAD.
The framework equips the UAD with continual learning capability through contrastively-learned prompts.
We conduct comprehensive experiments and set the benchmark on unsupervised continual anomaly detection and segmentation.
arXiv Detail & Related papers (2024-01-02T03:37:11Z) - SSB: Simple but Strong Baseline for Boosting Performance of Open-Set
Semi-Supervised Learning [106.46648817126984]
In this paper, we study the challenging and realistic open-set SSL setting.
The goal is to both correctly classify inliers and to detect outliers.
We find that inlier classification performance can be largely improved by incorporating high-confidence pseudo-labeled data.
arXiv Detail & Related papers (2023-11-17T15:14:40Z) - Quantile-based Maximum Likelihood Training for Outlier Detection [5.902139925693801]
We introduce a quantile-based maximum likelihood objective for learning the inlier distribution to improve the outlier separation during inference.
Our approach fits a normalizing flow to pre-trained discriminative features and detects the outliers according to the evaluated log-likelihood.
arXiv Detail & Related papers (2023-08-20T22:27:54Z) - ODIM: Outlier Detection via Likelihood of Under-Fitted Generative Models [4.956259629094216]
unsupervised outlier detection (UOD) problem refers to a task to identify inliers given training data which contain outliers as well as inliers.
We develop a new method called the outlier detection via the IM effect (ODIM)
Remarkably, the ODIM requires only a few updates, making it computationally efficient at least tens of times faster than other deep-learning-based algorithms.
arXiv Detail & Related papers (2023-01-11T01:02:27Z) - Comparison of Outlier Detection Techniques for Structured Data [2.2907341026741017]
An outlier is an observation or a data point that is far from rest of the data points in a given dataset.
It is seen that the removal of outliers from the training dataset before modeling can give better predictions.
The goal of this work is to highlight and compare some of the existing outlier detection techniques for the data scientists to use that information for outlier algorithm selection.
arXiv Detail & Related papers (2021-06-16T13:40:02Z) - Do We Really Need to Learn Representations from In-domain Data for
Outlier Detection? [6.445605125467574]
Methods based on the two-stage framework achieve state-of-the-art performance on this task.
We explore the possibility of avoiding the high cost of training a distinct representation for each outlier detection task.
In experiments, we demonstrate competitive or better performance on a variety of outlier detection benchmarks compared with previous two-stage methods.
arXiv Detail & Related papers (2021-05-19T17:30:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.