GraphAnoGAN: Detecting Anomalous Snapshots from Attributed Graphs
- URL: http://arxiv.org/abs/2106.15504v1
- Date: Tue, 29 Jun 2021 15:35:37 GMT
- Title: GraphAnoGAN: Detecting Anomalous Snapshots from Attributed Graphs
- Authors: Siddharth Bhatia, Yiwei Wang, Bryan Hooi, Tanmoy Chakraborty
- Abstract summary: We propose GraphAnoGAN, an anomalous snapshot ranking framework.
It consists of two core components -- generative and discriminative models.
Experiments on 4 real-world networks show that GraphAnoGAN outperforms 6 baselines with a significant margin.
- Score: 36.00861758441135
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Finding anomalous snapshots from a graph has garnered huge attention
recently. Existing studies address the problem using shallow learning
mechanisms such as subspace selection, ego-network, or community analysis.
These models do not take into account the multifaceted interactions between the
structure and attributes in the network. In this paper, we propose GraphAnoGAN,
an anomalous snapshot ranking framework, which consists of two core components
-- generative and discriminative models. Specifically, the generative model
learns to approximate the distribution of anomalous samples from the candidate
set of graph snapshots, and the discriminative model detects whether the
sampled snapshot is from the ground-truth or not. Experiments on 4 real-world
networks show that GraphAnoGAN outperforms 6 baselines with a significant
margin (28.29% and 22.01% higher precision and recall, respectively compared to
the best baseline, averaged across all datasets).
Related papers
- Generation is better than Modification: Combating High Class Homophily Variance in Graph Anomaly Detection [51.11833609431406]
Homophily distribution differences between different classes are significantly greater than those in homophilic and heterophilic graphs.
We introduce a new metric called Class Homophily Variance, which quantitatively describes this phenomenon.
To mitigate its impact, we propose a novel GNN model named Homophily Edge Generation Graph Neural Network (HedGe)
arXiv Detail & Related papers (2024-03-15T14:26:53Z) - Graph Out-of-Distribution Generalization with Controllable Data
Augmentation [51.17476258673232]
Graph Neural Network (GNN) has demonstrated extraordinary performance in classifying graph properties.
Due to the selection bias of training and testing data, distribution deviation is widespread.
We propose OOD calibration to measure the distribution deviation of virtual samples.
arXiv Detail & Related papers (2023-08-16T13:10:27Z) - Challenging the Myth of Graph Collaborative Filtering: a Reasoned and Reproducibility-driven Analysis [50.972595036856035]
We present a code that successfully replicates results from six popular and recent graph recommendation models.
We compare these graph models with traditional collaborative filtering models that historically performed well in offline evaluations.
By investigating the information flow from users' neighborhoods, we aim to identify which models are influenced by intrinsic features in the dataset structure.
arXiv Detail & Related papers (2023-08-01T09:31:44Z) - Generated Graph Detection [27.591612297045817]
Graph generative models become increasingly effective for data distribution approximation and data augmentation.
We propose the first framework to investigate a set of sophisticated models and their performance in four classification scenarios.
Our solution can sustain for a decent while to curb generated graph misuses.
arXiv Detail & Related papers (2023-06-13T13:18:04Z) - DAGAD: Data Augmentation for Graph Anomaly Detection [57.92471847260541]
This paper devises a novel Data Augmentation-based Graph Anomaly Detection (DAGAD) framework for attributed graphs.
A series of experiments on three datasets prove that DAGAD outperforms ten state-of-the-art baseline detectors concerning various mostly-used metrics.
arXiv Detail & Related papers (2022-10-18T11:28:21Z) - Graph Anomaly Detection with Unsupervised GNNs [19.772490600670363]
We design GLAM, an end-to-end graph-level anomaly detection model based on graph neural networks (GNNs)
We also propose a new pooling strategy for graph-level embedding, called MMD-pooling, that is geared toward detecting distribution anomalies which has not been considered before.
arXiv Detail & Related papers (2022-10-18T01:59:58Z) - Deep Graph-level Anomaly Detection by Glocal Knowledge Distillation [61.39364567221311]
Graph-level anomaly detection (GAD) describes the problem of detecting graphs that are abnormal in their structure and/or the features of their nodes.
One of the challenges in GAD is to devise graph representations that enable the detection of both locally- and globally-anomalous graphs.
We introduce a novel deep anomaly detection approach for GAD that learns rich global and local normal pattern information by joint random distillation of graph and node representations.
arXiv Detail & Related papers (2021-12-19T05:04:53Z) - Graph Classification by Mixture of Diverse Experts [67.33716357951235]
We present GraphDIVE, a framework leveraging mixture of diverse experts for imbalanced graph classification.
With a divide-and-conquer principle, GraphDIVE employs a gating network to partition an imbalanced graph dataset into several subsets.
Experiments on real-world imbalanced graph datasets demonstrate the effectiveness of GraphDIVE.
arXiv Detail & Related papers (2021-03-29T14:03:03Z) - Issues with Propagation Based Models for Graph-Level Outlier Detection [16.980621769406916]
Graph-Level Outlier Detection ( GLOD) is the task of identifying unusual graphs within a graph database.
This paper identifies and delves into a fundamental and intriguing issue with applying propagation based models to GLOD.
We find that ROC-AUC performance of the models change significantly depending on which class is down-sampled.
arXiv Detail & Related papers (2020-12-23T19:38:21Z) - Understanding Anomaly Detection with Deep Invertible Networks through
Hierarchies of Distributions and Features [4.25227087152716]
Convolutional networks learn similar low-level feature distributions when trained on any natural image dataset.
When the discriminative features between inliers and outliers are on a high-level, anomaly detection becomes particularly challenging.
We propose two methods to remove the negative impact of model bias and domain prior on detecting high-level differences.
arXiv Detail & Related papers (2020-06-18T20:56:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.