Improving Fairness in Graph Neural Networks via Mitigating Sensitive
Attribute Leakage
- URL: http://arxiv.org/abs/2206.03426v2
- Date: Thu, 9 Jun 2022 23:20:38 GMT
- Title: Improving Fairness in Graph Neural Networks via Mitigating Sensitive
Attribute Leakage
- Authors: Yu Wang, Yuying Zhao, Yushun Dong, Huiyuan Chen, Jundong Li, Tyler
Derr
- Abstract summary: Graph Neural Networks (GNNs) have shown great power in learning node representations on graphs.
GNNs may inherit historical prejudices from training data, leading to discriminatory bias in predictions.
We propose Fair View Graph Neural Network (FairVGNN) to generate fair views of features by automatically identifying and masking sensitive-correlated features.
- Score: 35.810534649478576
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph Neural Networks (GNNs) have shown great power in learning node
representations on graphs. However, they may inherit historical prejudices from
training data, leading to discriminatory bias in predictions. Although some
work has developed fair GNNs, most of them directly borrow fair representation
learning techniques from non-graph domains without considering the potential
problem of sensitive attribute leakage caused by feature propagation in GNNs.
However, we empirically observe that feature propagation could vary the
correlation of previously innocuous non-sensitive features to the sensitive
ones. This can be viewed as a leakage of sensitive information which could
further exacerbate discrimination in predictions. Thus, we design two feature
masking strategies according to feature correlations to highlight the
importance of considering feature propagation and correlation variation in
alleviating discrimination. Motivated by our analysis, we propose Fair View
Graph Neural Network (FairVGNN) to generate fair views of features by
automatically identifying and masking sensitive-correlated features considering
correlation variation after feature propagation. Given the learned fair views,
we adaptively clamp weights of the encoder to avoid using sensitive-related
features. Experiments on real-world datasets demonstrate that FairVGNN enjoys a
better trade-off between model utility and fairness. Our code is publicly
available at https://github.com/YuWVandy/FairVGNN.
Related papers
- One Fits All: Learning Fair Graph Neural Networks for Various Sensitive Attributes [40.57757706386367]
We propose a graph fairness framework based on invariant learning, namely FairINV.
FairINV incorporates sensitive attribute partition and trains fair GNNs by eliminating spurious correlations between the label and various sensitive attributes.
Experimental results on several real-world datasets demonstrate that FairINV significantly outperforms state-of-the-art fairness approaches.
arXiv Detail & Related papers (2024-06-19T13:30:17Z) - Fair Graph Neural Network with Supervised Contrastive Regularization [12.666235467177131]
We propose a novel model for training fairness-aware Graph Neural Networks (GNNs)
Our approach integrates Supervised Contrastive Loss and Environmental Loss to enhance both accuracy and fairness.
arXiv Detail & Related papers (2024-04-09T07:49:05Z) - FairSIN: Achieving Fairness in Graph Neural Networks through Sensitive Information Neutralization [40.78945336901361]
We argue that filtering-based strategies may also filter out some non-sensitive feature information, leading to a sub-optimal trade-off between predictive performance and fairness.
We unveil an innovative neutralization-based paradigm, where additional Fairness-facilitating Features (F3) are incorporated into node features or representations before message passing.
We name our method as FairSIN, and present three implementation variants from both data-centric and model-centric perspectives.
arXiv Detail & Related papers (2024-03-19T06:22:58Z) - Endowing Pre-trained Graph Models with Provable Fairness [49.8431177748876]
We propose a novel adapter-tuning framework that endows pre-trained graph models with provable fairness (called GraphPAR)
Specifically, we design a sensitive semantic augmenter on node representations, to extend the node representations with different sensitive attribute semantics for each node.
With GraphPAR, we quantify whether the fairness of each node is provable, i.e., predictions are always fair within a certain range of sensitive attribute semantics.
arXiv Detail & Related papers (2024-02-19T14:16:08Z) - Graph Out-of-Distribution Generalization via Causal Intervention [69.70137479660113]
We introduce a conceptually simple yet principled approach for training robust graph neural networks (GNNs) under node-level distribution shifts.
Our method resorts to a new learning objective derived from causal inference that coordinates an environment estimator and a mixture-of-expert GNN predictor.
Our model can effectively enhance generalization with various types of distribution shifts and yield up to 27.4% accuracy improvement over state-of-the-arts on graph OOD generalization benchmarks.
arXiv Detail & Related papers (2024-02-18T07:49:22Z) - Chasing Fairness in Graphs: A GNN Architecture Perspective [73.43111851492593]
We propose textsfFair textsfMessage textsfPassing (FMP) designed within a unified optimization framework for graph neural networks (GNNs)
In FMP, the aggregation is first adopted to utilize neighbors' information and then the bias mitigation step explicitly pushes demographic group node presentation centers together.
Experiments on node classification tasks demonstrate that the proposed FMP outperforms several baselines in terms of fairness and accuracy on three real-world datasets.
arXiv Detail & Related papers (2023-12-19T18:00:15Z) - Towards Fair Graph Neural Networks via Graph Counterfactual [38.721295940809135]
Graph neural networks (GNNs) have shown great ability in representation (GNNs) learning on graphs, facilitating various tasks.
Recent works show that GNNs tend to inherit and amplify the bias from training data, causing concerns of the adoption of GNNs in high-stake scenarios.
We propose a novel framework CAF, which can select counterfactuals from training data to avoid non-realistic counterfactuals.
arXiv Detail & Related papers (2023-07-10T23:28:03Z) - Fairness-Aware Graph Neural Networks: A Survey [53.41838868516936]
Graph Neural Networks (GNNs) have become increasingly important due to their representational power and state-of-the-art predictive performance.
GNNs suffer from fairness issues that arise as a result of the underlying graph data and the fundamental aggregation mechanism.
In this article, we examine and categorize fairness techniques for improving the fairness of GNNs.
arXiv Detail & Related papers (2023-07-08T08:09:06Z) - Generalizing Graph Neural Networks on Out-Of-Distribution Graphs [51.33152272781324]
Graph Neural Networks (GNNs) are proposed without considering the distribution shifts between training and testing graphs.
In such a setting, GNNs tend to exploit subtle statistical correlations existing in the training set for predictions, even though it is a spurious correlation.
We propose a general causal representation framework, called StableGNN, to eliminate the impact of spurious correlations.
arXiv Detail & Related papers (2021-11-20T18:57:18Z) - Discriminability of Single-Layer Graph Neural Networks [172.5042368548269]
Graph neural networks (GNNs) have exhibited promising performance on a wide range of problems.
We focus on the property of discriminability and establish conditions under which the inclusion of pointwise nonlinearities to a stable graph filter bank leads to an increased discriminative capacity for high-eigenvalue content.
arXiv Detail & Related papers (2020-10-17T18:52:34Z) - Say No to the Discrimination: Learning Fair Graph Neural Networks with
Limited Sensitive Attribute Information [37.90997236795843]
Graph neural networks (GNNs) have shown great power in modeling graph structured data.
GNNs may make predictions biased on protected sensitive attributes, e.g., skin color and gender.
We propose FairGNN to eliminate the bias of GNNs whilst maintaining high node classification accuracy.
arXiv Detail & Related papers (2020-09-03T05:17:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.