On Graph Neural Network Fairness in the Presence of Heterophilous
Neighborhoods
- URL: http://arxiv.org/abs/2207.04376v2
- Date: Mon, 14 Nov 2022 22:39:39 GMT
- Title: On Graph Neural Network Fairness in the Presence of Heterophilous
Neighborhoods
- Authors: Donald Loveland, Jiong Zhu, Mark Heimann, Ben Fish, Michael T. Schaub,
Danai Koutra
- Abstract summary: We study the task of node classification for graph neural networks (GNNs)
We establish a connection between group fairness, as measured by statistical parity and equal opportunity, and local assortativity.
We show that by adopting heterophilous GNN designs, group fairness in locally heterophilous neighborhoods can be improved by up to 25% over homophilous designs in real and synthetic datasets.
- Score: 12.531373572440787
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We study the task of node classification for graph neural networks (GNNs) and
establish a connection between group fairness, as measured by statistical
parity and equal opportunity, and local assortativity, i.e., the tendency of
linked nodes to have similar attributes. Such assortativity is often induced by
homophily, the tendency for nodes of similar properties to connect. Homophily
can be common in social networks where systemic factors have forced individuals
into communities which share a sensitive attribute. Through synthetic graphs,
we study the interplay between locally occurring homophily and fair
predictions, finding that not all node neighborhoods are equal in this respect
-- neighborhoods dominated by one category of a sensitive attribute often
struggle to obtain fair treatment, especially in the case of diverging local
class and sensitive attribute homophily. After determining that a relationship
between local homophily and fairness exists, we investigate if the issue of
unfairness can be associated to the design of the applied GNN model. We show
that by adopting heterophilous GNN designs capable of handling disassortative
group labels, group fairness in locally heterophilous neighborhoods can be
improved by up to 25% over homophilous designs in real and synthetic datasets.
Related papers
- Unveiling the Impact of Local Homophily on GNN Fairness: In-Depth Analysis and New Benchmarks [12.077386474432124]
Graph Neural Networks (GNNs) often struggle to generalize when graphs exhibit both homophily (same-class connections) and heterophily (different-class connections)
This issue poses a risk in user-centric applications where underrepresented homophily levels are present.
We move beyond global homophily and explore how local homophily levels can lead to unfair predictions.
arXiv Detail & Related papers (2024-10-05T21:21:40Z) - The Heterophilic Graph Learning Handbook: Benchmarks, Models, Theoretical Analysis, Applications and Challenges [101.83124435649358]
Homophily principle, ie nodes with the same labels or similar attributes are more likely to be connected.
Recent work has identified a non-trivial set of datasets where GNN's performance compared to the NN's is not satisfactory.
arXiv Detail & Related papers (2024-07-12T18:04:32Z) - Federated Graph Semantic and Structural Learning [54.97668931176513]
This paper reveals that local client distortion is brought by both node-level semantics and graph-level structure.
We postulate that a well-structural graph neural network possesses similarity for neighbors due to the inherent adjacency relationships.
We transform the adjacency relationships into the similarity distribution and leverage the global model to distill the relation knowledge into the local model.
arXiv Detail & Related papers (2024-06-27T07:08:28Z) - Heterophilous Distribution Propagation for Graph Neural Networks [23.897535976924722]
We propose heterophilous distribution propagation (HDP) for graph neural networks.
Instead of aggregating information from all neighborhoods, HDP adaptively separates the neighbors into homophilous and heterphilous parts.
We conduct extensive experiments on 9 benchmark datasets with different levels of homophily.
arXiv Detail & Related papers (2024-05-31T06:40:56Z) - Graph Out-of-Distribution Generalization via Causal Intervention [69.70137479660113]
We introduce a conceptually simple yet principled approach for training robust graph neural networks (GNNs) under node-level distribution shifts.
Our method resorts to a new learning objective derived from causal inference that coordinates an environment estimator and a mixture-of-expert GNN predictor.
Our model can effectively enhance generalization with various types of distribution shifts and yield up to 27.4% accuracy improvement over state-of-the-arts on graph OOD generalization benchmarks.
arXiv Detail & Related papers (2024-02-18T07:49:22Z) - Demystifying Structural Disparity in Graph Neural Networks: Can One Size
Fit All? [61.35457647107439]
Most real-world homophilic and heterophilic graphs are comprised of a mixture of nodes in both homophilic and heterophilic structural patterns.
We provide evidence that Graph Neural Networks(GNNs) on node classification typically perform admirably on homophilic nodes.
We then propose a rigorous, non-i.i.d PAC-Bayesian generalization bound for GNNs, revealing reasons for the performance disparity.
arXiv Detail & Related papers (2023-06-02T07:46:20Z) - LSGNN: Towards General Graph Neural Network in Node Classification by
Local Similarity [59.41119013018377]
We propose to use the local similarity (LocalSim) to learn node-level weighted fusion, which can also serve as a plug-and-play module.
For better fusion, we propose a novel and efficient Initial Residual Difference Connection (IRDC) to extract more informative multi-hop information.
Our proposed method, namely Local Similarity Graph Neural Network (LSGNN), can offer comparable or superior state-of-the-art performance on both homophilic and heterophilic graphs.
arXiv Detail & Related papers (2023-05-07T09:06:11Z) - Heterophily-Aware Graph Attention Network [42.640057865981156]
Graph Neural Networks (GNNs) have shown remarkable success in graph representation learning.
Existing heterophilic GNNs tend to ignore the modeling of heterophily of each edge, which is also a vital part in tackling the heterophily problem.
We propose a novel Heterophily-Aware Graph Attention Network (HA-GAT) by fully exploring and utilizing the local distribution as the underlying heterophily.
arXiv Detail & Related papers (2023-02-07T03:21:55Z) - Revisiting Heterophily For Graph Neural Networks [42.41238892727136]
Graph Neural Networks (GNNs) extend basic Neural Networks (NNs) by using graph structures based on the relational inductive bias (homophily assumption)
Recent work has identified a non-trivial set of datasets where their performance compared to NNs is not satisfactory.
arXiv Detail & Related papers (2022-10-14T08:00:26Z) - Is Homophily a Necessity for Graph Neural Networks? [50.959340355849896]
Graph neural networks (GNNs) have shown great prowess in learning representations suitable for numerous graph-based machine learning tasks.
GNNs are widely believed to work well due to the homophily assumption ("like attracts like"), and fail to generalize to heterophilous graphs where dissimilar nodes connect.
Recent works design new architectures to overcome such heterophily-related limitations, citing poor baseline performance and new architecture improvements on a few heterophilous graph benchmark datasets as evidence for this notion.
In our experiments, we empirically find that standard graph convolutional networks (GCNs) can actually achieve better performance than
arXiv Detail & Related papers (2021-06-11T02:44:00Z) - Beyond Homophily in Graph Neural Networks: Current Limitations and
Effective Designs [28.77753005139331]
We investigate the representation power of graph neural networks in a semi-supervised node classification task under heterophily or low homophily.
Many popular GNNs fail to generalize to this setting, and are even outperformed by models that ignore the graph structure.
We identify a set of key designs that boost learning from the graph structure under heterophily.
arXiv Detail & Related papers (2020-06-20T02:05:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.