Cold Brew: Distilling Graph Node Representations with Incomplete or
Missing Neighborhoods
- URL: http://arxiv.org/abs/2111.04840v2
- Date: Wed, 10 Nov 2021 01:40:04 GMT
- Title: Cold Brew: Distilling Graph Node Representations with Incomplete or
Missing Neighborhoods
- Authors: Wenqing Zheng, Edward W Huang, Nikhil Rao, Sumeet Katariya, Zhangyang
Wang and Karthik Subbian
- Abstract summary: We introduce feature-contribution ratio (FCR) to study the viability of using inductive GNNs to solve the Strict Cold Start (SCS) problem.
We experimentally show FCR disentangles the contributions of various components of graph datasets and demonstrate the superior performance of Cold Brew.
- Score: 69.13371028670153
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Graph Neural Networks (GNNs) have achieved state of the art performance in
node classification, regression, and recommendation tasks. GNNs work well when
high-quality and rich connectivity structure is available. However, this
requirement is not satisfied in many real world graphs where the node degrees
have power-law distributions as many nodes have either fewer or noisy
connections. The extreme case of this situation is a node may have no neighbors
at all, called Strict Cold Start (SCS) scenario. This forces the prediction
models to rely completely on the node's input features. We propose Cold Brew to
address the SCS and noisy neighbor setting compared to pointwise and other
graph-based models via a distillation approach. We introduce
feature-contribution ratio (FCR), a metric to study the viability of using
inductive GNNs to solve the SCS problem and to select the best architecture for
SCS generalization. We experimentally show FCR disentangles the contributions
of various components of graph datasets and demonstrate the superior
performance of Cold Brew on several public benchmarks and proprietary
e-commerce datasets. The source code for our approach is available at:
https://github.com/amazon-research/gnn-tail-generalization.
Related papers
- G-SPARC: SPectral ARchitectures tackling the Cold-start problem in Graph learning [3.870455775654713]
We propose G-SPARC, a novel framework addressing cold-start nodes.
By utilizing a key idea of transitioning from graph representation to spectral representation, our approach is generalizable to cold-start nodes.
Our method outperforms existing models on cold-start nodes across various tasks like node classification, node clustering, and link prediction.
arXiv Detail & Related papers (2024-11-03T11:39:09Z) - Spectral Greedy Coresets for Graph Neural Networks [61.24300262316091]
The ubiquity of large-scale graphs in node-classification tasks hinders the real-world applications of Graph Neural Networks (GNNs)
This paper studies graph coresets for GNNs and avoids the interdependence issue by selecting ego-graphs based on their spectral embeddings.
Our spectral greedy graph coreset (SGGC) scales to graphs with millions of nodes, obviates the need for model pre-training, and applies to low-homophily graphs.
arXiv Detail & Related papers (2024-05-27T17:52:12Z) - LSGNN: Towards General Graph Neural Network in Node Classification by
Local Similarity [59.41119013018377]
We propose to use the local similarity (LocalSim) to learn node-level weighted fusion, which can also serve as a plug-and-play module.
For better fusion, we propose a novel and efficient Initial Residual Difference Connection (IRDC) to extract more informative multi-hop information.
Our proposed method, namely Local Similarity Graph Neural Network (LSGNN), can offer comparable or superior state-of-the-art performance on both homophilic and heterophilic graphs.
arXiv Detail & Related papers (2023-05-07T09:06:11Z) - Exploiting Neighbor Effect: Conv-Agnostic GNNs Framework for Graphs with
Heterophily [58.76759997223951]
We propose a new metric based on von Neumann entropy to re-examine the heterophily problem of GNNs.
We also propose a Conv-Agnostic GNN framework (CAGNNs) to enhance the performance of most GNNs on heterophily datasets.
arXiv Detail & Related papers (2022-03-19T14:26:43Z) - Inferential SIR-GN: Scalable Graph Representation Learning [0.4699313647907615]
Graph representation learning methods generate numerical vector representations for the nodes in a network.
In this work, we propose Inferential SIR-GN, a model which is pre-trained on random graphs, then computes node representations rapidly.
We demonstrate that the model is able to capture node's structural role information, and show excellent performance at node and graph classification tasks, on unseen networks.
arXiv Detail & Related papers (2021-11-08T20:56:37Z) - Position-based Hash Embeddings For Scaling Graph Neural Networks [8.87527266373087]
Graph Neural Networks (GNNs) compute node representations by taking into account the topology of the node's ego-network and the features of the ego-network's nodes.
When the nodes do not have high-quality features, GNNs learn an embedding layer to compute node embeddings and use them as input features.
To reduce the memory associated with this embedding layer, hashing-based approaches, commonly used in applications like NLP and recommender systems, can potentially be used.
We present approaches that take advantage of the nodes' position in the graph to dramatically reduce the memory required.
arXiv Detail & Related papers (2021-08-31T22:42:25Z) - Scalable Graph Neural Networks for Heterogeneous Graphs [12.44278942365518]
Graph neural networks (GNNs) are a popular class of parametric model for learning over graph-structured data.
Recent work has argued that GNNs primarily use the graph for feature smoothing, and have shown competitive results on benchmark tasks.
In this work, we ask whether these results can be extended to heterogeneous graphs, which encode multiple types of relationship between different entities.
arXiv Detail & Related papers (2020-11-19T06:03:35Z) - Node Similarity Preserving Graph Convolutional Networks [51.520749924844054]
Graph Neural Networks (GNNs) explore the graph structure and node features by aggregating and transforming information within node neighborhoods.
We propose SimP-GCN that can effectively and efficiently preserve node similarity while exploiting graph structure.
We validate the effectiveness of SimP-GCN on seven benchmark datasets including three assortative and four disassorative graphs.
arXiv Detail & Related papers (2020-11-19T04:18:01Z) - Bilinear Graph Neural Network with Neighbor Interactions [106.80781016591577]
Graph Neural Network (GNN) is a powerful model to learn representations and make predictions on graph data.
We propose a new graph convolution operator, which augments the weighted sum with pairwise interactions of the representations of neighbor nodes.
We term this framework as Bilinear Graph Neural Network (BGNN), which improves GNN representation ability with bilinear interactions between neighbor nodes.
arXiv Detail & Related papers (2020-02-10T06:43:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.