The Impact of Global Structural Information in Graph Neural Networks
Applications
- URL: http://arxiv.org/abs/2006.03814v2
- Date: Wed, 15 Dec 2021 17:37:31 GMT
- Title: The Impact of Global Structural Information in Graph Neural Networks
Applications
- Authors: Davide Buffelli, Fabio Vandin
- Abstract summary: Graph Neural Networks (GNNs) rely on the graph structure to define an aggregation strategy.
A known limitation of GNNs is that, as the number of layers increases, information gets smoothed and squashed.
We give access to global information to several GNN models and observe the impact it has on downstream performance.
- Score: 5.629161809575013
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph Neural Networks (GNNs) rely on the graph structure to define an
aggregation strategy where each node updates its representation by combining
information from its neighbours. A known limitation of GNNs is that, as the
number of layers increases, information gets smoothed and squashed and node
embeddings become indistinguishable, negatively affecting performance.
Therefore, practical GNN models employ few layers and only leverage the graph
structure in terms of limited, small neighbourhoods around each node.
Inevitably, practical GNNs do not capture information depending on the global
structure of the graph. While there have been several works studying the
limitations and expressivity of GNNs, the question of whether practical
applications on graph structured data require global structural knowledge or
not, remains unanswered. In this work, we empirically address this question by
giving access to global information to several GNN models, and observing the
impact it has on downstream performance. Our results show that global
information can in fact provide significant benefits for common graph-related
tasks. We further identify a novel regularization strategy that leads to an
average accuracy improvement of more than 5% on all considered tasks.
Related papers
- MDS-GNN: A Mutual Dual-Stream Graph Neural Network on Graphs with Incomplete Features and Structure [8.00268216176428]
Graph Neural Networks (GNNs) have emerged as powerful tools for analyzing and learning representations from graph-structured data.
A crucial prerequisite for the outstanding performance of GNNs is the availability of complete graph information.
This study proposes a mutual dual-stream graph neural network (MDS-GNN) which implements a mutual benefit learning between features and structure.
arXiv Detail & Related papers (2024-08-09T03:42:56Z) - Unveiling Global Interactive Patterns across Graphs: Towards Interpretable Graph Neural Networks [31.29616732552006]
Graph Neural Networks (GNNs) have emerged as a prominent framework for graph mining.
This paper proposes a novel intrinsically interpretable scheme for graph classification.
Global Interactive Pattern (GIP) learning introduces learnable global interactive patterns to explicitly interpret decisions.
arXiv Detail & Related papers (2024-07-02T06:31:13Z) - Harnessing Collective Structure Knowledge in Data Augmentation for Graph Neural Networks [25.12261412297796]
Graph neural networks (GNNs) have achieved state-of-the-art performance in graph representation learning.
We propose a novel approach, namely collective structure knowledge-augmented graph neural network (CoS-GNN)
arXiv Detail & Related papers (2024-05-17T08:50:00Z) - DGNN: Decoupled Graph Neural Networks with Structural Consistency
between Attribute and Graph Embedding Representations [62.04558318166396]
Graph neural networks (GNNs) demonstrate a robust capability for representation learning on graphs with complex structures.
A novel GNNs framework, dubbed Decoupled Graph Neural Networks (DGNN), is introduced to obtain a more comprehensive embedding representation of nodes.
Experimental results conducted on several graph benchmark datasets verify DGNN's superiority in node classification task.
arXiv Detail & Related papers (2024-01-28T06:43:13Z) - Learning Strong Graph Neural Networks with Weak Information [64.64996100343602]
We develop a principled approach to the problem of graph learning with weak information (GLWI)
We propose D$2$PT, a dual-channel GNN framework that performs long-range information propagation on the input graph with incomplete structure, but also on a global graph that encodes global semantic similarities.
arXiv Detail & Related papers (2023-05-29T04:51:09Z) - DEGREE: Decomposition Based Explanation For Graph Neural Networks [55.38873296761104]
We propose DEGREE to provide a faithful explanation for GNN predictions.
By decomposing the information generation and aggregation mechanism of GNNs, DEGREE allows tracking the contributions of specific components of the input graph to the final prediction.
We also design a subgraph level interpretation algorithm to reveal complex interactions between graph nodes that are overlooked by previous methods.
arXiv Detail & Related papers (2023-05-22T10:29:52Z) - Discovering the Representation Bottleneck of Graph Neural Networks from
Multi-order Interactions [51.597480162777074]
Graph neural networks (GNNs) rely on the message passing paradigm to propagate node features and build interactions.
Recent works point out that different graph learning tasks require different ranges of interactions between nodes.
We study two common graph construction methods in scientific domains, i.e., emphK-nearest neighbor (KNN) graphs and emphfully-connected (FC) graphs.
arXiv Detail & Related papers (2022-05-15T11:38:14Z) - Topological Graph Neural Networks [14.349152231293928]
We present TOGL, a novel layer that incorporates global topological information of a graph using persistent homology.
Augmenting GNNs with our layer leads to beneficial predictive performance, both on synthetic data sets and on real-world data.
arXiv Detail & Related papers (2021-02-15T20:27:56Z) - Learning to Drop: Robust Graph Neural Network via Topological Denoising [50.81722989898142]
We propose PTDNet, a parameterized topological denoising network, to improve the robustness and generalization performance of Graph Neural Networks (GNNs)
PTDNet prunes task-irrelevant edges by penalizing the number of edges in the sparsified graph with parameterized networks.
We show that PTDNet can improve the performance of GNNs significantly and the performance gain becomes larger for more noisy datasets.
arXiv Detail & Related papers (2020-11-13T18:53:21Z) - From Local Structures to Size Generalization in Graph Neural Networks [53.3202754533658]
Graph neural networks (GNNs) can process graphs of different sizes.
Their ability to generalize across sizes, specifically from small to large graphs, is still not well understood.
arXiv Detail & Related papers (2020-10-17T19:36:54Z) - Incomplete Graph Representation and Learning via Partial Graph Neural
Networks [7.227805463462352]
In many applications, graph may be coming in an incomplete form where attributes of graph nodes are partially unknown/missing.
Existing GNNs are generally designed on complete graphs which can not deal with attribute-incomplete graph data directly.
We develop a novel partial aggregation based GNNs, named Partial Graph Neural Networks (PaGNNs) for attribute-incomplete graph representation and learning.
arXiv Detail & Related papers (2020-03-23T08:29:59Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.