Debiased Graph Neural Networks with Agnostic Label Selection Bias
- URL: http://arxiv.org/abs/2201.07708v1
- Date: Wed, 19 Jan 2022 16:50:29 GMT
- Title: Debiased Graph Neural Networks with Agnostic Label Selection Bias
- Authors: Shaohua Fan, Xiao Wang, Chuan Shi, Kun Kuang, Nian Liu, Bai Wang
- Abstract summary: Most existing Graph Neural Networks (GNNs) are proposed without considering the selection bias in data.
We propose a novel Debiased Graph Neural Networks (DGNN) with a differentiated decorrelation regularizer.
Our proposed model outperforms the state-of-the-art methods and DGNN is a flexible framework to enhance existing GNNs.
- Score: 59.61301255860836
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Most existing Graph Neural Networks (GNNs) are proposed without considering
the selection bias in data, i.e., the inconsistent distribution between the
training set with test set. In reality, the test data is not even available
during the training process, making selection bias agnostic. Training GNNs with
biased selected nodes leads to significant parameter estimation bias and
greatly impacts the generalization ability on test nodes. In this paper, we
first present an experimental investigation, which clearly shows that the
selection bias drastically hinders the generalization ability of GNNs, and
theoretically prove that the selection bias will cause the biased estimation on
GNN parameters. Then to remove the bias in GNN estimation, we propose a novel
Debiased Graph Neural Networks (DGNN) with a differentiated decorrelation
regularizer. The differentiated decorrelation regularizer estimates a sample
weight for each labeled node such that the spurious correlation of learned
embeddings could be eliminated. We analyze the regularizer in causal view and
it motivates us to differentiate the weights of the variables based on their
contribution on the confounding bias. Then, these sample weights are used for
reweighting GNNs to eliminate the estimation bias, thus help to improve the
stability of prediction on unknown test nodes. Comprehensive experiments are
conducted on several challenging graph datasets with two kinds of label
selection biases. The results well verify that our proposed model outperforms
the state-of-the-art methods and DGNN is a flexible framework to enhance
existing GNNs.
Related papers
- Learning to Reweight for Graph Neural Network [63.978102332612906]
Graph Neural Networks (GNNs) show promising results for graph tasks.
Existing GNNs' generalization ability will degrade when there exist distribution shifts between testing and training graph data.
We propose a novel nonlinear graph decorrelation method, which can substantially improve the out-of-distribution generalization ability.
arXiv Detail & Related papers (2023-12-19T12:25:10Z) - Causality and Independence Enhancement for Biased Node Classification [56.38828085943763]
We propose a novel Causality and Independence Enhancement (CIE) framework, applicable to various graph neural networks (GNNs)
Our approach estimates causal and spurious features at the node representation level and mitigates the influence of spurious correlations.
Our approach CIE not only significantly enhances the performance of GNNs but outperforms state-of-the-art debiased node classification methods.
arXiv Detail & Related papers (2023-10-14T13:56:24Z) - Interpreting Unfairness in Graph Neural Networks via Training Node
Attribution [46.384034587689136]
We study a novel problem of interpreting GNN unfairness through attributing it to the influence of training nodes.
Specifically, we propose a novel strategy named Probabilistic Distribution Disparity (PDD) to measure the bias exhibited in GNNs.
We verify the validity of PDD and the effectiveness of influence estimation through experiments on real-world datasets.
arXiv Detail & Related papers (2022-11-25T21:52:30Z) - Generalizing Graph Neural Networks on Out-Of-Distribution Graphs [51.33152272781324]
Graph Neural Networks (GNNs) are proposed without considering the distribution shifts between training and testing graphs.
In such a setting, GNNs tend to exploit subtle statistical correlations existing in the training set for predictions, even though it is a spurious correlation.
We propose a general causal representation framework, called StableGNN, to eliminate the impact of spurious correlations.
arXiv Detail & Related papers (2021-11-20T18:57:18Z) - Stable Prediction on Graphs with Agnostic Distribution Shift [105.12836224149633]
Graph neural networks (GNNs) have been shown to be effective on various graph tasks with randomly separated training and testing data.
In real applications, however, the distribution of training graph might be different from that of the test one.
We propose a novel stable prediction framework for GNNs, which permits both locally and globally stable learning and prediction on graphs.
arXiv Detail & Related papers (2021-10-08T02:45:47Z) - EDITS: Modeling and Mitigating Data Bias for Graph Neural Networks [29.974829042502375]
We develop a framework named EDITS to mitigate the bias in attributed networks.
EDITS works in a model-agnostic manner, which means that it is independent of the specific GNNs applied for downstream tasks.
arXiv Detail & Related papers (2021-08-11T14:07:01Z) - Shift-Robust GNNs: Overcoming the Limitations of Localized Graph
Training data [52.771780951404565]
Shift-Robust GNN (SR-GNN) is designed to account for distributional differences between biased training data and the graph's true inference distribution.
We show that SR-GNN outperforms other GNN baselines by accuracy, eliminating at least (40%) of the negative effects introduced by biased training data.
arXiv Detail & Related papers (2021-08-02T18:00:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.