Impact Of Missing Data Imputation On The Fairness And Accuracy Of Graph
Node Classifiers
- URL: http://arxiv.org/abs/2211.00783v1
- Date: Tue, 1 Nov 2022 23:16:36 GMT
- Title: Impact Of Missing Data Imputation On The Fairness And Accuracy Of Graph
Node Classifiers
- Authors: Haris Mansoor, Sarwan Ali, Shafiq Alam, Muhammad Asad Khan, Umair ul
Hassan, Imdadullah Khan
- Abstract summary: We analyze the effect on fairness in the context of graph data (node attributes) imputation using different embedding and neural network methods.
Our results provide valuable insights into graph data fairness and how to handle missingness in graphs efficiently.
- Score: 0.19573380763700707
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Analysis of the fairness of machine learning (ML) algorithms recently
attracted many researchers' interest. Most ML methods show bias toward
protected groups, which limits the applicability of ML models in many
applications like crime rate prediction etc. Since the data may have missing
values which, if not appropriately handled, are known to further harmfully
affect fairness. Many imputation methods are proposed to deal with missing
data. However, the effect of missing data imputation on fairness is not studied
well. In this paper, we analyze the effect on fairness in the context of graph
data (node attributes) imputation using different embedding and neural network
methods. Extensive experiments on six datasets demonstrate severe fairness
issues in missing data imputation under graph node classification. We also find
that the choice of the imputation method affects both fairness and accuracy.
Our results provide valuable insights into graph data fairness and how to
handle missingness in graphs efficiently. This work also provides directions
regarding theoretical studies on fairness in graph data.
Related papers
- Is it Still Fair? A Comparative Evaluation of Fairness Algorithms through the Lens of Covariate Drift [17.498879317113385]
We study data distributional drift and its impact on fairness algorithms and metrics.
In several cases, data distributional drift can lead to serious deterioration of fairness in so-called fair models.
Emanating from our findings, we synthesize several policy implications of data distributional drift on fairness algorithms.
arXiv Detail & Related papers (2024-09-19T03:18:12Z) - Fair GLASSO: Estimating Fair Graphical Models with Unbiased Statistical Behavior [31.92791228859847]
Many real-world models exhibit unfair discriminatory behavior due to biases in data.
We introduce fairness for graphical models in the form of two bias metrics to promote balance in statistical similarities.
We present Fair GLASSO, a regularized graphical lasso approach to obtain sparse Gaussian precision matrices.
arXiv Detail & Related papers (2024-06-13T18:07:04Z) - Graph Fairness Learning under Distribution Shifts [33.9878682279549]
Graph neural networks (GNNs) have achieved remarkable performance on graph-structured data.
GNNs may inherit prejudice from the training data and make discriminatory predictions based on sensitive attributes, such as gender and race.
We propose a graph generator to produce numerous graphs with significant bias and under different distances.
arXiv Detail & Related papers (2024-01-30T06:51:24Z) - GraphGuard: Detecting and Counteracting Training Data Misuse in Graph
Neural Networks [69.97213941893351]
The emergence of Graph Neural Networks (GNNs) in graph data analysis has raised critical concerns about data misuse during model training.
Existing methodologies address either data misuse detection or mitigation, and are primarily designed for local GNN models.
This paper introduces a pioneering approach called GraphGuard, to tackle these challenges.
arXiv Detail & Related papers (2023-12-13T02:59:37Z) - Deceptive Fairness Attacks on Graphs via Meta Learning [102.53029537886314]
We study deceptive fairness attacks on graphs to answer the question: How can we achieve poisoning attacks on a graph learning model to exacerbate the bias deceptively?
We propose a meta learning-based framework named FATE to attack various fairness definitions and graph learning models.
We conduct extensive experimental evaluations on real-world datasets in the task of semi-supervised node classification.
arXiv Detail & Related papers (2023-10-24T09:10:14Z) - Fair Attribute Completion on Graph with Missing Attributes [14.950261239035882]
We propose FairAC, a fair attribute completion method, to complement missing information and learn fair node embeddings for graphs with missing attributes.
We show that our method achieves better fairness performance with less sacrifice in accuracy, compared with the state-of-the-art methods of fair graph learning.
arXiv Detail & Related papers (2023-02-25T04:12:30Z) - D-BIAS: A Causality-Based Human-in-the-Loop System for Tackling
Algorithmic Bias [57.87117733071416]
We propose D-BIAS, a visual interactive tool that embodies human-in-the-loop AI approach for auditing and mitigating social biases.
A user can detect the presence of bias against a group by identifying unfair causal relationships in the causal network.
For each interaction, say weakening/deleting a biased causal edge, the system uses a novel method to simulate a new (debiased) dataset.
arXiv Detail & Related papers (2022-08-10T03:41:48Z) - Counterfactual Fairness with Partially Known Causal Graph [85.15766086381352]
This paper proposes a general method to achieve the notion of counterfactual fairness when the true causal graph is unknown.
We find that counterfactual fairness can be achieved as if the true causal graph were fully known, when specific background knowledge is provided.
arXiv Detail & Related papers (2022-05-27T13:40:50Z) - Distributionally Robust Semi-Supervised Learning Over Graphs [68.29280230284712]
Semi-supervised learning (SSL) over graph-structured data emerges in many network science applications.
To efficiently manage learning over graphs, variants of graph neural networks (GNNs) have been developed recently.
Despite their success in practice, most of existing methods are unable to handle graphs with uncertain nodal attributes.
Challenges also arise due to distributional uncertainties associated with data acquired by noisy measurements.
A distributionally robust learning framework is developed, where the objective is to train models that exhibit quantifiable robustness against perturbations.
arXiv Detail & Related papers (2021-10-20T14:23:54Z) - Can Active Learning Preemptively Mitigate Fairness Issues? [66.84854430781097]
dataset bias is one of the prevailing causes of unfairness in machine learning.
We study whether models trained with uncertainty-based ALs are fairer in their decisions with respect to a protected class.
We also explore the interaction of algorithmic fairness methods such as gradient reversal (GRAD) and BALD.
arXiv Detail & Related papers (2021-04-14T14:20:22Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.