Confidence May Cheat: Self-Training on Graph Neural Networks under
Distribution Shift
- URL: http://arxiv.org/abs/2201.11349v1
- Date: Thu, 27 Jan 2022 07:12:27 GMT
- Title: Confidence May Cheat: Self-Training on Graph Neural Networks under
Distribution Shift
- Authors: Hongrui Liu, Binbin Hu, Xiao Wang, Chuan Shi, Zhiqiang Zhang, Jun Zhou
- Abstract summary: Self-training methods have been widely adopted on graphs by labeling high-confidence unlabeled nodes and then adding them to the training step.
We propose a novel Distribution Recovered Graph Self-Training framework (DR- GST), which could recover the distribution of the original labeled dataset.
Both our theoretical analysis and extensive experiments on five benchmark datasets demonstrate the effectiveness of the proposed DR- GST.
- Score: 39.73304203101909
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph Convolutional Networks (GCNs) have recently attracted vast interest and
achieved state-of-the-art performance on graphs, but its success could
typically hinge on careful training with amounts of expensive and
time-consuming labeled data. To alleviate labeled data scarcity, self-training
methods have been widely adopted on graphs by labeling high-confidence
unlabeled nodes and then adding them to the training step. In this line, we
empirically make a thorough study for current self-training methods on graphs.
Surprisingly, we find that high-confidence unlabeled nodes are not always
useful, and even introduce the distribution shift issue between the original
labeled dataset and the augmented dataset by self-training, severely hindering
the capability of self-training on graphs. To this end, in this paper, we
propose a novel Distribution Recovered Graph Self-Training framework (DR-GST),
which could recover the distribution of the original labeled dataset.
Specifically, we first prove the equality of loss function in self-training
framework under the distribution shift case and the population distribution if
each pseudo-labeled node is weighted by a proper coefficient. Considering the
intractability of the coefficient, we then propose to replace the coefficient
with the information gain after observing the same changing trend between them,
where information gain is respectively estimated via both dropout variational
inference and dropedge variational inference in DR-GST. However, such a
weighted loss function will enlarge the impact of incorrect pseudo labels. As a
result, we apply the loss correction method to improve the quality of pseudo
labels. Both our theoretical analysis and extensive experiments on five
benchmark datasets demonstrate the effectiveness of the proposed DR-GST, as
well as each well-designed component in DR-GST.
Related papers
- Degree Distribution based Spiking Graph Networks for Domain Adaptation [17.924123705983792]
Spiking Graph Networks (SGNs) have garnered significant attraction from both researchers and industry due to their ability to address energy consumption challenges in graph classification.
We first propose the domain adaptation problem in SGNs, and introduce a novel framework named Degree-aware Spiking Graph Domain Adaptation for Classification.
The proposed DeSGDA addresses the spiking graph domain adaptation problem by three aspects: node degree-aware personalized spiking representation, adversarial feature distribution alignment, and pseudo-label distillation.
arXiv Detail & Related papers (2024-10-09T13:45:54Z) - Distribution Consistency based Self-Training for Graph Neural Networks
with Sparse Labels [33.89511660654271]
Few-shot node classification poses a significant challenge for Graph Neural Networks (GNNs)
Self-training has emerged as a widely popular framework to leverage the abundance of unlabeled data.
We propose a novel Distribution-Consistent Graph Self-Training framework to identify pseudo-labeled nodes that are both informative and capable of redeeming the distribution discrepancy.
arXiv Detail & Related papers (2024-01-18T22:07:48Z) - CONVERT:Contrastive Graph Clustering with Reliable Augmentation [110.46658439733106]
We propose a novel CONtrastiVe Graph ClustEring network with Reliable AugmenTation (CONVERT)
In our method, the data augmentations are processed by the proposed reversible perturb-recover network.
To further guarantee the reliability of semantics, a novel semantic loss is presented to constrain the network.
arXiv Detail & Related papers (2023-08-17T13:07:09Z) - Graph Out-of-Distribution Generalization with Controllable Data
Augmentation [51.17476258673232]
Graph Neural Network (GNN) has demonstrated extraordinary performance in classifying graph properties.
Due to the selection bias of training and testing data, distribution deviation is widespread.
We propose OOD calibration to measure the distribution deviation of virtual samples.
arXiv Detail & Related papers (2023-08-16T13:10:27Z) - All Points Matter: Entropy-Regularized Distribution Alignment for
Weakly-supervised 3D Segmentation [67.30502812804271]
Pseudo-labels are widely employed in weakly supervised 3D segmentation tasks where only sparse ground-truth labels are available for learning.
We propose a novel learning strategy to regularize the generated pseudo-labels and effectively narrow the gaps between pseudo-labels and model predictions.
arXiv Detail & Related papers (2023-05-25T08:19:31Z) - OOD-GNN: Out-of-Distribution Generalized Graph Neural Network [73.67049248445277]
Graph neural networks (GNNs) have achieved impressive performance when testing and training graph data come from identical distribution.
Existing GNNs lack out-of-distribution generalization abilities so that their performance substantially degrades when there exist distribution shifts between testing and training graph data.
We propose an out-of-distribution generalized graph neural network (OOD-GNN) for achieving satisfactory performance on unseen testing graphs that have different distributions with training graphs.
arXiv Detail & Related papers (2021-12-07T16:29:10Z) - Stable Prediction on Graphs with Agnostic Distribution Shift [105.12836224149633]
Graph neural networks (GNNs) have been shown to be effective on various graph tasks with randomly separated training and testing data.
In real applications, however, the distribution of training graph might be different from that of the test one.
We propose a novel stable prediction framework for GNNs, which permits both locally and globally stable learning and prediction on graphs.
arXiv Detail & Related papers (2021-10-08T02:45:47Z) - A Graph Data Augmentation Strategy with Entropy Preserving [11.886325179121226]
We introduce a novel graph entropy definition as a quantitative index to evaluate feature information among a graph.
Under considerations of preserving graph entropy, we propose an effective strategy to generate training data using a perturbed mechanism.
Our proposed approach significantly enhances the robustness and generalization ability of GCNs during the training process.
arXiv Detail & Related papers (2021-07-13T12:58:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.