ACGAN-GNNExplainer: Auxiliary Conditional Generative Explainer for Graph
Neural Networks
- URL: http://arxiv.org/abs/2309.16918v2
- Date: Wed, 11 Oct 2023 00:06:34 GMT
- Title: ACGAN-GNNExplainer: Auxiliary Conditional Generative Explainer for Graph
Neural Networks
- Authors: Yiqiao Li, Jianlong Zhou, Yifei Dong, Niusha Shafiabady, Fang Chen
- Abstract summary: Graph neural networks (GNNs) have proven their efficacy in a variety of real-world applications, but their underlying mechanisms remain a mystery.
To address this challenge and enable reliable decision-making, many GNN explainers have been proposed in recent years.
We introduce Auxiliary Generative Adrative Network (ACGAN) into the field of GNN explanation and propose a new GNN explainer dubbedemphACGANGNNExplainer.
- Score: 7.077341403454516
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph neural networks (GNNs) have proven their efficacy in a variety of
real-world applications, but their underlying mechanisms remain a mystery. To
address this challenge and enable reliable decision-making, many GNN explainers
have been proposed in recent years. However, these methods often encounter
limitations, including their dependence on specific instances, lack of
generalizability to unseen graphs, producing potentially invalid explanations,
and yielding inadequate fidelity. To overcome these limitations, we, in this
paper, introduce the Auxiliary Classifier Generative Adversarial Network
(ACGAN) into the field of GNN explanation and propose a new GNN explainer
dubbed~\emph{ACGAN-GNNExplainer}. Our approach leverages a generator to produce
explanations for the original input graphs while incorporating a discriminator
to oversee the generation process, ensuring explanation fidelity and improving
accuracy. Experimental evaluations conducted on both synthetic and real-world
graph datasets demonstrate the superiority of our proposed method compared to
other existing GNN explainers.
Related papers
- Global Graph Counterfactual Explanation: A Subgraph Mapping Approach [54.42907350881448]
Graph Neural Networks (GNNs) have been widely deployed in various real-world applications.
Counterfactual explanation aims to find minimum perturbations on input graphs that change the GNN predictions.
We propose GlobalGCE, a novel global-level graph counterfactual explanation method.
arXiv Detail & Related papers (2024-10-25T21:39:05Z) - DEGREE: Decomposition Based Explanation For Graph Neural Networks [55.38873296761104]
We propose DEGREE to provide a faithful explanation for GNN predictions.
By decomposing the information generation and aggregation mechanism of GNNs, DEGREE allows tracking the contributions of specific components of the input graph to the final prediction.
We also design a subgraph level interpretation algorithm to reveal complex interactions between graph nodes that are overlooked by previous methods.
arXiv Detail & Related papers (2023-05-22T10:29:52Z) - GANExplainer: GAN-based Graph Neural Networks Explainer [5.641321839562139]
It is critical to explain why graph neural network (GNN) makes particular predictions for them to be believed in many applications.
We propose GANExplainer, based on Generative Adversarial Network (GAN) architecture.
GANExplainer improves explanation accuracy by up to 35% compared to its alternatives.
arXiv Detail & Related papers (2022-12-30T23:11:24Z) - Task-Agnostic Graph Explanations [50.17442349253348]
Graph Neural Networks (GNNs) have emerged as powerful tools to encode graph structured data.
Existing learning-based GNN explanation approaches are task-specific in training.
We propose a Task-Agnostic GNN Explainer (TAGE) trained under self-supervision with no knowledge of downstream tasks.
arXiv Detail & Related papers (2022-02-16T21:11:47Z) - Jointly Attacking Graph Neural Network and its Explanations [50.231829335996814]
Graph Neural Networks (GNNs) have boosted the performance for many graph-related tasks.
Recent studies have shown that GNNs are highly vulnerable to adversarial attacks, where adversaries can mislead the GNNs' prediction by modifying graphs.
We propose a novel attack framework (GEAttack) which can attack both a GNN model and its explanations by simultaneously exploiting their vulnerabilities.
arXiv Detail & Related papers (2021-08-07T07:44:33Z) - GCExplainer: Human-in-the-Loop Concept-based Explanations for Graph
Neural Networks [0.3441021278275805]
GCExplainer is an unsupervised approach for post-hoc discovery and extraction of global concept-based explanations for graph neural networks (GNNs)
We demonstrate the success of our technique on five node classification datasets and two graph classification datasets, showing that we are able to discover and extract high-quality concept representations by putting the human in the loop.
arXiv Detail & Related papers (2021-07-25T20:52:48Z) - SEEN: Sharpening Explanations for Graph Neural Networks using
Explanations from Neighborhoods [0.0]
We propose a method to improve the explanation quality of node classification tasks through aggregation of auxiliary explanations.
Applying SEEN does not require modification of a graph and can be used with diverse explainability techniques.
Experiments on matching motif-participating nodes from a given graph show great improvement in explanation accuracy of up to 12.71%.
arXiv Detail & Related papers (2021-06-16T03:04:46Z) - Zorro: Valid, Sparse, and Stable Explanations in Graph Neural Networks [6.004582130591279]
We find that the previous explanation generation approaches that maximize the mutual information between the label distribution produced by the GNN model and the explanation to be restrictive.
Specifically, existing approaches do not enforce explanations to be predictive, sparse, or robust to input perturbations.
We propose a novel approach Zorro based on the principles from rate-distortion theory that uses a simple procedure to optimize for fidelity.
arXiv Detail & Related papers (2021-05-18T15:53:09Z) - Parameterized Explainer for Graph Neural Network [49.79917262156429]
We propose PGExplainer, a parameterized explainer for Graph Neural Networks (GNNs)
Compared to the existing work, PGExplainer has better generalization ability and can be utilized in an inductive setting easily.
Experiments on both synthetic and real-life datasets show highly competitive performance with up to 24.7% relative improvement in AUC on explaining graph classification.
arXiv Detail & Related papers (2020-11-09T17:15:03Z) - XGNN: Towards Model-Level Explanations of Graph Neural Networks [113.51160387804484]
Graphs neural networks (GNNs) learn node features by aggregating and combining neighbor information.
GNNs are mostly treated as black-boxes and lack human intelligible explanations.
We propose a novel approach, known as XGNN, to interpret GNNs at the model-level.
arXiv Detail & Related papers (2020-06-03T23:52:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.