AdaGCN:Adaptive Boosting Algorithm for Graph Convolutional Networks on
Imbalanced Node Classification
- URL: http://arxiv.org/abs/2105.11625v1
- Date: Tue, 25 May 2021 02:43:31 GMT
- Title: AdaGCN:Adaptive Boosting Algorithm for Graph Convolutional Networks on
Imbalanced Node Classification
- Authors: S. Shi, Kai Qiao, Shuai Yang, L. Wang, J. Chen and Bin Yan
- Abstract summary: We propose an ensemble model called AdaGCN, which uses a Graph Convolutional Network (GCN) as the base estimator during adaptive boosting.
Our model also improves state-of-the-art baselines on all of the challenging node classification tasks we consider.
- Score: 10.72543417177307
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Graph Neural Network (GNN) has achieved remarkable success in graph data
representation. However, the previous work only considered the ideal balanced
dataset, and the practical imbalanced dataset was rarely considered, which, on
the contrary, is of more significance for the application of GNN. Traditional
methods such as resampling, reweighting and synthetic samples that deal with
imbalanced datasets are no longer applicable in GNN. Ensemble models can handle
imbalanced datasets better compared with single estimator. Besides, ensemble
learning can achieve higher estimation accuracy and has better reliability
compared with the single estimator. In this paper, we propose an ensemble model
called AdaGCN, which uses a Graph Convolutional Network (GCN) as the base
estimator during adaptive boosting. In AdaGCN, a higher weight will be set for
the training samples that are not properly classified by the previous
classifier, and transfer learning is used to reduce computational cost and
increase fitting capability. Experiments show that the AdaGCN model we proposed
achieves better performance than GCN, GraphSAGE, GAT, N-GCN and the most of
advanced reweighting and resampling methods on synthetic imbalanced datasets,
with an average improvement of 4.3%. Our model also improves state-of-the-art
baselines on all of the challenging node classification tasks we consider:
Cora, Citeseer, Pubmed, and NELL.
Related papers
- Graph Mining under Data scarcity [6.229055041065048]
We propose an Uncertainty Estimator framework that can be applied on top of any generic Graph Neural Networks (GNNs)
We train these models under the classic episodic learning paradigm in the $n$-way, $k$-shot fashion, in an end-to-end setting.
Our method outperforms the baselines, which demonstrates the efficacy of the Uncertainty Estimator for Few-shot node classification on graphs with a GNN.
arXiv Detail & Related papers (2024-06-07T10:50:03Z) - Class-Balanced and Reinforced Active Learning on Graphs [13.239043161351482]
Graph neural networks (GNNs) have demonstrated significant success in various applications, such as node classification, link prediction, and graph classification.
Active learning for GNNs aims to query the valuable samples from the unlabeled data for annotation to maximize the GNNs' performance at a lower cost.
Most existing algorithms for reinforced active learning in GNNs may lead to a highly imbalanced class distribution, especially in highly skewed class scenarios.
We propose a novel class-balanced and reinforced active learning framework for GNNs, namely, GCBR. It learns an optimal policy to acquire class-balanced and informative nodes
arXiv Detail & Related papers (2024-02-15T16:37:14Z) - Learning to Reweight for Graph Neural Network [63.978102332612906]
Graph Neural Networks (GNNs) show promising results for graph tasks.
Existing GNNs' generalization ability will degrade when there exist distribution shifts between testing and training graph data.
We propose a novel nonlinear graph decorrelation method, which can substantially improve the out-of-distribution generalization ability.
arXiv Detail & Related papers (2023-12-19T12:25:10Z) - GNNEvaluator: Evaluating GNN Performance On Unseen Graphs Without Labels [81.93520935479984]
We study a new problem, GNN model evaluation, that aims to assess the performance of a specific GNN model trained on labeled and observed graphs.
We propose a two-stage GNN model evaluation framework, including (1) DiscGraph set construction and (2) GNNEvaluator training and inference.
Under the effective training supervision from the DiscGraph set, GNNEvaluator learns to precisely estimate node classification accuracy of the to-be-evaluated GNN model.
arXiv Detail & Related papers (2023-10-23T05:51:59Z) - Label Deconvolution for Node Representation Learning on Large-scale
Attributed Graphs against Learning Bias [75.44877675117749]
We propose an efficient label regularization technique, namely Label Deconvolution (LD), to alleviate the learning bias by a novel and highly scalable approximation to the inverse mapping of GNNs.
Experiments demonstrate LD significantly outperforms state-of-the-art methods on Open Graph datasets Benchmark.
arXiv Detail & Related papers (2023-09-26T13:09:43Z) - Comprehensive Graph Gradual Pruning for Sparse Training in Graph Neural
Networks [52.566735716983956]
We propose a graph gradual pruning framework termed CGP to dynamically prune GNNs.
Unlike LTH-based methods, the proposed CGP approach requires no re-training, which significantly reduces the computation costs.
Our proposed strategy greatly improves both training and inference efficiency while matching or even exceeding the accuracy of existing methods.
arXiv Detail & Related papers (2022-07-18T14:23:31Z) - Shift-Robust GNNs: Overcoming the Limitations of Localized Graph
Training data [52.771780951404565]
Shift-Robust GNN (SR-GNN) is designed to account for distributional differences between biased training data and the graph's true inference distribution.
We show that SR-GNN outperforms other GNN baselines by accuracy, eliminating at least (40%) of the negative effects introduced by biased training data.
arXiv Detail & Related papers (2021-08-02T18:00:38Z) - Automated Graph Learning via Population Based Self-Tuning GCN [45.28411311903644]
Graph convolutional network (GCN) and its variants have been successfully applied to a broad range of tasks.
Traditional GCN models suffer from the issues of overfitting and oversmoothing.
Recent techniques like DropEdge could alleviate these issues and thus enable the development of deep GCN.
arXiv Detail & Related papers (2021-07-09T23:05:21Z) - Bayesian Graph Neural Networks with Adaptive Connection Sampling [62.51689735630133]
We propose a unified framework for adaptive connection sampling in graph neural networks (GNNs)
The proposed framework not only alleviates over-smoothing and over-fitting tendencies of deep GNNs, but also enables learning with uncertainty in graph analytic tasks with GNNs.
arXiv Detail & Related papers (2020-06-07T07:06:35Z) - Self-Enhanced GNN: Improving Graph Neural Networks Using Model Outputs [20.197085398581397]
Graph neural networks (GNNs) have received much attention recently because of their excellent performance on graph-based tasks.
We propose self-enhanced GNN (SEG), which improves the quality of the input data using the outputs of existing GNN models.
SEG consistently improves the performance of well-known GNN models such as GCN, GAT and SGC across different datasets.
arXiv Detail & Related papers (2020-02-18T12:27:16Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.