Are we really making much progress in unsupervised graph outlier
detection? Revisiting the problem with new insight and superior method
- URL: http://arxiv.org/abs/2210.12941v1
- Date: Mon, 24 Oct 2022 04:09:35 GMT
- Title: Are we really making much progress in unsupervised graph outlier
detection? Revisiting the problem with new insight and superior method
- Authors: Yihong Huang, Liping Wang, Fan Zhang, Xuemin Lin
- Abstract summary: UNOD focuses on detecting two kinds of typical outliers in graphs: the structural outlier and the contextual outlier.
We find that the most widely-used outlier injection approach has a serious data leakage issue.
We propose a new framework, Variance-based Graph Outlier Detection (VGOD), which combines our variance-based model and attribute reconstruction model to detect outliers in a balanced way.
- Score: 36.72922385614812
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: A large number of studies on Graph Outlier Detection (GOD) have emerged in
recent years due to its wide applications, in which Unsupervised Node Outlier
Detection (UNOD) on attributed networks is an important area. UNOD focuses on
detecting two kinds of typical outliers in graphs: the structural outlier and
the contextual outlier. Most existing works conduct the experiments based on
the datasets with injected outliers. However, we find that the most widely-used
outlier injection approach has a serious data leakage issue. By only utilizing
such data leakage, a simple approach can achieve the state-of-the-art
performance in detecting outliers. In addition, we observe that most existing
algorithms have performance drops with varied injection settings. The other
major issue is on balanced detection performance between the two types of
outliers, which has not been considered by existing studies. In this paper, we
analyze the cause of the data leakage issue in depth since the injection
approach is a building block to advance UNOD. Moreover, we devise a novel
variance-based model to detect structural outliers, which is more robust to
different injection settings. On top of this, we propose a new framework,
Variance-based Graph Outlier Detection (VGOD), which combines our
variance-based model and attribute reconstruction model to detect outliers in a
balanced way. Finally, we conduct extensive experiments to demonstrate the
effectiveness and the efficiency of VGOD. The results on 5 real-world datasets
validate that VGOD achieves not only the best performance in detecting outliers
but also a balanced detection performance between structural and contextual
outliers.
Related papers
- HGOE: Hybrid External and Internal Graph Outlier Exposure for Graph Out-of-Distribution Detection [78.47008997035158]
Graph data exhibits greater diversity but lower robustness to perturbations, complicating the integration of outliers.
We propose the introduction of textbfHybrid External and Internal textbfGraph textbfOutlier textbfExposure (HGOE) to improve graph OOD detection performance.
arXiv Detail & Related papers (2024-07-31T16:55:18Z) - Data Augmentation for Supervised Graph Outlier Detection with Latent Diffusion Models [39.33024157496401]
We introduce GODM, a novel data augmentation for mitigating class imbalance in supervised graph outlier detection with latent Diffusion Models.
Our proposed method consists of three key components: (1) Variantioanl maps the heterogeneous information inherent within the graph data into a unified latent space, (2) Graph Generator synthesizes graph data that are statistically similar to real outliers from latent space, and (3) Latent Diffusion Model learns the latent space distribution of real organic data by iterative denoising.
arXiv Detail & Related papers (2023-12-29T16:50:40Z) - Diversified Outlier Exposure for Out-of-Distribution Detection via
Informative Extrapolation [110.34982764201689]
Out-of-distribution (OOD) detection is important for deploying reliable machine learning models on real-world applications.
Recent advances in outlier exposure have shown promising results on OOD detection via fine-tuning model with informatively sampled auxiliary outliers.
We propose a novel framework, namely, Diversified Outlier Exposure (DivOE), for effective OOD detection via informative extrapolation based on the given auxiliary outliers.
arXiv Detail & Related papers (2023-10-21T07:16:09Z) - ODIM: Outlier Detection via Likelihood of Under-Fitted Generative Models [4.956259629094216]
unsupervised outlier detection (UOD) problem refers to a task to identify inliers given training data which contain outliers as well as inliers.
We develop a new method called the outlier detection via the IM effect (ODIM)
Remarkably, the ODIM requires only a few updates, making it computationally efficient at least tens of times faster than other deep-learning-based algorithms.
arXiv Detail & Related papers (2023-01-11T01:02:27Z) - DEGAN: Time Series Anomaly Detection using Generative Adversarial
Network Discriminators and Density Estimation [0.0]
We have proposed an unsupervised Generative Adversarial Network (GAN)-based anomaly detection framework, DEGAN.
It relies solely on normal time series data as input to train a well-configured discriminator (D) into a standalone anomaly predictor.
arXiv Detail & Related papers (2022-10-05T04:32:12Z) - An Outlier Exposure Approach to Improve Visual Anomaly Detection
Performance for Mobile Robots [76.36017224414523]
We consider the problem of building visual anomaly detection systems for mobile robots.
Standard anomaly detection models are trained using large datasets composed only of non-anomalous data.
We tackle the problem of exploiting these data to improve the performance of a Real-NVP anomaly detection model.
arXiv Detail & Related papers (2022-09-20T15:18:13Z) - Efficient remedies for outlier detection with variational autoencoders [8.80692072928023]
Likelihoods computed by deep generative models are a candidate metric for outlier detection with unlabeled data.
We show that a theoretically-grounded correction readily ameliorates a key bias with VAE likelihood estimates.
We also show that the variance of the likelihoods computed over an ensemble of VAEs also enables robust outlier detection.
arXiv Detail & Related papers (2021-08-19T16:00:58Z) - Homophily Outlier Detection in Non-IID Categorical Data [43.51919113927003]
This work introduces a novel outlier detection framework and its two instances to identify outliers in categorical data.
It first defines and incorporates distribution-sensitive outlier factors and their interdependence into a value-value graph-based representation.
The learned value outlierness allows for either direct outlier detection or outlying feature selection.
arXiv Detail & Related papers (2021-03-21T23:29:33Z) - SUOD: Accelerating Large-Scale Unsupervised Heterogeneous Outlier
Detection [63.253850875265115]
Outlier detection (OD) is a key machine learning (ML) task for identifying abnormal objects from general samples.
We propose a modular acceleration system, called SUOD, to address it.
arXiv Detail & Related papers (2020-03-11T00:22:50Z) - Generalized ODIN: Detecting Out-of-distribution Image without Learning
from Out-of-distribution Data [87.61504710345528]
We propose two strategies for freeing a neural network from tuning with OoD data, while improving its OoD detection performance.
We specifically propose to decompose confidence scoring as well as a modified input pre-processing method.
Our further analysis on a larger scale image dataset shows that the two types of distribution shifts, specifically semantic shift and non-semantic shift, present a significant difference.
arXiv Detail & Related papers (2020-02-26T04:18:25Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.