Toward the Analysis of Graph Neural Networks
- URL: http://arxiv.org/abs/2201.00115v1
- Date: Sat, 1 Jan 2022 04:59:49 GMT
- Title: Toward the Analysis of Graph Neural Networks
- Authors: Thanh-Dat Nguyen, Thanh Le-Cong, ThanhVu H. Nguyen, Xuan-Bach D. Le,
Quyet-Thang Huynh
- Abstract summary: Graph Neural Networks (GNNs) have emerged as a robust framework for graph-structured data analysis.
This paper proposes an approach to analyze GNNs by converting them into Feed Forward Neural Networks (FFNNs) and reusing existing FFNNs analyses.
- Score: 1.0412114420493723
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: Graph Neural Networks (GNNs) have recently emerged as a robust framework for
graph-structured data. They have been applied to many problems such as
knowledge graph analysis, social networks recommendation, and even Covid19
detection and vaccine developments. However, unlike other deep neural networks
such as Feed Forward Neural Networks (FFNNs), few analyses such as verification
and property inferences exist, potentially due to dynamic behaviors of GNNs,
which can take arbitrary graphs as input, whereas FFNNs which only take fixed
size numerical vectors as inputs.
This paper proposes an approach to analyze GNNs by converting them into FFNNs
and reusing existing FFNNs analyses. We discuss various designs to ensure the
scalability and accuracy of the conversions. We illustrate our method on a
study case of node classification. We believe that our approach opens new
research directions for understanding and analyzing GNNs.
Related papers
- Uncertainty in Graph Neural Networks: A Survey [50.63474656037679]
Graph Neural Networks (GNNs) have been extensively used in various real-world applications.
However, the predictive uncertainty of GNNs stemming from diverse sources can lead to unstable and erroneous predictions.
This survey aims to provide a comprehensive overview of the GNNs from the perspective of uncertainty.
arXiv Detail & Related papers (2024-03-11T21:54:52Z) - Information Flow in Graph Neural Networks: A Clinical Triage Use Case [49.86931948849343]
Graph Neural Networks (GNNs) have gained popularity in healthcare and other domains due to their ability to process multi-modal and multi-relational graphs.
We investigate how the flow of embedding information within GNNs affects the prediction of links in Knowledge Graphs (KGs)
Our results demonstrate that incorporating domain knowledge into the GNN connectivity leads to better performance than using the same connectivity as the KG or allowing unconstrained embedding propagation.
arXiv Detail & Related papers (2023-09-12T09:18:12Z) - DEGREE: Decomposition Based Explanation For Graph Neural Networks [55.38873296761104]
We propose DEGREE to provide a faithful explanation for GNN predictions.
By decomposing the information generation and aggregation mechanism of GNNs, DEGREE allows tracking the contributions of specific components of the input graph to the final prediction.
We also design a subgraph level interpretation algorithm to reveal complex interactions between graph nodes that are overlooked by previous methods.
arXiv Detail & Related papers (2023-05-22T10:29:52Z) - A Survey of Explainable Graph Neural Networks: Taxonomy and Evaluation
Metrics [8.795591344648294]
We focus on explainable graph neural networks and categorize them based on the use of explainable methods.
We provide the common performance metrics for GNNs explanations and point out several future research directions.
arXiv Detail & Related papers (2022-07-26T01:45:54Z) - Explainability in Graph Neural Networks: An Experimental Survey [12.440636971075977]
Graph neural networks (GNNs) have been extensively developed for graph representation learning.
GNNs suffer from the black-box problem as people cannot understand the mechanism underlying them.
Several GNN explainability methods have been proposed to explain the decisions made by GNNs.
arXiv Detail & Related papers (2022-03-17T11:25:41Z) - Edge-Level Explanations for Graph Neural Networks by Extending
Explainability Methods for Convolutional Neural Networks [33.20913249848369]
Graph Neural Networks (GNNs) are deep learning models that take graph data as inputs, and they are applied to various tasks such as traffic prediction and molecular property prediction.
We extend explainability methods for CNNs, such as Local Interpretable Model-Agnostic Explanations (LIME), Gradient-Based Saliency Maps, and Gradient-Weighted Class Activation Mapping (Grad-CAM) to GNNs.
The experimental results indicate that the LIME-based approach is the most efficient explainability method for multiple tasks in the real-world situation, outperforming even the state-of-the
arXiv Detail & Related papers (2021-11-01T06:27:29Z) - GNNLens: A Visual Analytics Approach for Prediction Error Diagnosis of
Graph Neural Networks [42.222552078920216]
Graph Neural Networks (GNNs) aim to extend deep learning techniques to graph data.
GNNs behave like a black box with their details hidden from model developers and users.
It is therefore difficult to diagnose possible errors of GNNs.
This paper fills the research gap with an interactive visual analysis tool, GNNLens, to assist model developers and users in understanding and analyzing GNNs.
arXiv Detail & Related papers (2020-11-22T16:09:08Z) - A Unified View on Graph Neural Networks as Graph Signal Denoising [49.980783124401555]
Graph Neural Networks (GNNs) have risen to prominence in learning representations for graph structured data.
In this work, we establish mathematically that the aggregation processes in a group of representative GNN models can be regarded as solving a graph denoising problem.
We instantiate a novel GNN model, ADA-UGNN, derived from UGNN, to handle graphs with adaptive smoothness across nodes.
arXiv Detail & Related papers (2020-10-05T04:57:18Z) - Graph Neural Networks: Architectures, Stability and Transferability [176.3960927323358]
Graph Neural Networks (GNNs) are information processing architectures for signals supported on graphs.
They are generalizations of convolutional neural networks (CNNs) in which individual layers contain banks of graph convolutional filters.
arXiv Detail & Related papers (2020-08-04T18:57:36Z) - Adversarial Attacks and Defenses on Graphs: A Review, A Tool and
Empirical Studies [73.39668293190019]
Adversary attacks can be easily fooled by small perturbation on the input.
Graph Neural Networks (GNNs) have been demonstrated to inherit this vulnerability.
In this survey, we categorize existing attacks and defenses, and review the corresponding state-of-the-art methods.
arXiv Detail & Related papers (2020-03-02T04:32:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.