Stochastic Subgraph Neighborhood Pooling for Subgraph Classification
- URL: http://arxiv.org/abs/2304.08556v1
- Date: Mon, 17 Apr 2023 18:49:18 GMT
- Title: Stochastic Subgraph Neighborhood Pooling for Subgraph Classification
- Authors: Shweta Ann Jacob, Paul Louis and Amirali Salehi-Abari
- Abstract summary: Subgraph Neighborhood Pooling (SSNP) jointly aggregates the subgraph and its neighborhood information without any computationally expensive operations such as labeling tricks.
Our experiments demonstrate that our models outperform current state-of-the-art methods (with a margin of up to 2%) while being up to 3X faster in training.
- Score: 2.1270496914042996
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Subgraph classification is an emerging field in graph representation learning
where the task is to classify a group of nodes (i.e., a subgraph) within a
graph. Subgraph classification has applications such as predicting the cellular
function of a group of proteins or identifying rare diseases given a collection
of phenotypes. Graph neural networks (GNNs) are the de facto solution for node,
link, and graph-level tasks but fail to perform well on subgraph classification
tasks. Even GNNs tailored for graph classification are not directly
transferable to subgraph classification as they ignore the external topology of
the subgraph, thus failing to capture how the subgraph is located within the
larger graph. The current state-of-the-art models for subgraph classification
address this shortcoming through either labeling tricks or multiple
message-passing channels, both of which impose a computation burden and are not
scalable to large graphs. To address the scalability issue while maintaining
generalization, we propose Stochastic Subgraph Neighborhood Pooling (SSNP),
which jointly aggregates the subgraph and its neighborhood (i.e., external
topology) information without any computationally expensive operations such as
labeling tricks. To improve scalability and generalization further, we also
propose a simple data augmentation pre-processing step for SSNP that creates
multiple sparse views of the subgraph neighborhood. We show that our model is
more expressive than GNNs without labeling tricks. Our extensive experiments
demonstrate that our models outperform current state-of-the-art methods (with a
margin of up to 2%) while being up to 3X faster in training.
Related papers
- A Flexible, Equivariant Framework for Subgraph GNNs via Graph Products and Graph Coarsening [18.688057947275112]
Subgraph Graph Neural Networks (Subgraph GNNs) enhance the expressivity of message-passing GNNs by representing graphs as sets of subgraphs.
Previous approaches suggested processing only subsets of subgraphs, selected either randomly or via learnable sampling.
This paper introduces a new Subgraph GNNs framework to address these issues.
arXiv Detail & Related papers (2024-06-13T16:29:06Z) - SPGNN: Recognizing Salient Subgraph Patterns via Enhanced Graph Convolution and Pooling [25.555741218526464]
Graph neural networks (GNNs) have revolutionized the field of machine learning on non-Euclidean data such as graphs and networks.
We propose a concatenation-based graph convolution mechanism that injectively updates node representations.
We also design a novel graph pooling module, called WL-SortPool, to learn important subgraph patterns in a deep-learning manner.
arXiv Detail & Related papers (2024-04-21T13:11:59Z) - NodeFormer: A Scalable Graph Structure Learning Transformer for Node
Classification [70.51126383984555]
We introduce a novel all-pair message passing scheme for efficiently propagating node signals between arbitrary nodes.
The efficient computation is enabled by a kernerlized Gumbel-Softmax operator.
Experiments demonstrate the promising efficacy of the method in various tasks including node classification on graphs.
arXiv Detail & Related papers (2023-06-14T09:21:15Z) - Learning to Count Isomorphisms with Graph Neural Networks [16.455234748896157]
Subgraph isomorphism counting is an important problem on graphs.
In this paper, we propose a novel graph neural network (GNN) called Count-GNN for subgraph isomorphism counting.
arXiv Detail & Related papers (2023-02-07T05:32:11Z) - Position-Aware Subgraph Neural Networks with Data-Efficient Learning [15.58680146160525]
We propose a Position-Aware Data-Efficient Learning framework for subgraph neural networks called PADEL.
Specifically, we propose a novel node position encoding method that is anchor-free, and design a new generative subgraph augmentation method based on a diffused variational subgraph autoencoder.
arXiv Detail & Related papers (2022-11-01T16:34:42Z) - Personalized Subgraph Federated Learning [56.52903162729729]
We introduce a new subgraph FL problem, personalized subgraph FL, which focuses on the joint improvement of the interrelated local GNNs.
We propose a novel framework, FEDerated Personalized sUBgraph learning (FED-PUB), to tackle it.
We validate our FED-PUB for its subgraph FL performance on six datasets, considering both non-overlapping and overlapping subgraphs.
arXiv Detail & Related papers (2022-06-21T09:02:53Z) - Neural Graph Matching for Pre-training Graph Neural Networks [72.32801428070749]
Graph neural networks (GNNs) have been shown powerful capacity at modeling structural data.
We present a novel Graph Matching based GNN Pre-Training framework, called GMPT.
The proposed method can be applied to fully self-supervised pre-training and coarse-grained supervised pre-training.
arXiv Detail & Related papers (2022-03-03T09:53:53Z) - On Explainability of Graph Neural Networks via Subgraph Explorations [48.56936527708657]
We propose a novel method, known as SubgraphX, to explain graph neural networks (GNNs)
Our work represents the first attempt to explain GNNs via identifying subgraphs explicitly.
arXiv Detail & Related papers (2021-02-09T22:12:26Z) - Scalable Graph Neural Networks for Heterogeneous Graphs [12.44278942365518]
Graph neural networks (GNNs) are a popular class of parametric model for learning over graph-structured data.
Recent work has argued that GNNs primarily use the graph for feature smoothing, and have shown competitive results on benchmark tasks.
In this work, we ask whether these results can be extended to heterogeneous graphs, which encode multiple types of relationship between different entities.
arXiv Detail & Related papers (2020-11-19T06:03:35Z) - Line Graph Neural Networks for Link Prediction [71.00689542259052]
We consider the graph link prediction task, which is a classic graph analytical problem with many real-world applications.
In this formalism, a link prediction problem is converted to a graph classification task.
We propose to seek a radically different and novel path by making use of the line graphs in graph theory.
In particular, each node in a line graph corresponds to a unique edge in the original graph. Therefore, link prediction problems in the original graph can be equivalently solved as a node classification problem in its corresponding line graph, instead of a graph classification task.
arXiv Detail & Related papers (2020-10-20T05:54:31Z) - Graph Pooling with Node Proximity for Hierarchical Representation
Learning [80.62181998314547]
We propose a novel graph pooling strategy that leverages node proximity to improve the hierarchical representation learning of graph data with their multi-hop topology.
Results show that the proposed graph pooling strategy is able to achieve state-of-the-art performance on a collection of public graph classification benchmark datasets.
arXiv Detail & Related papers (2020-06-19T13:09:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.