Heterophily-Aware Fair Recommendation using Graph Convolutional Networks
- URL: http://arxiv.org/abs/2402.03365v3
- Date: Wed, 19 Feb 2025 14:15:17 GMT
- Title: Heterophily-Aware Fair Recommendation using Graph Convolutional Networks
- Authors: Nemat Gholinejad, Mostafa Haghir Chehreghani,
- Abstract summary: We propose a fair GNN-based recommender system, called HetroFair, to improve item-side fairness.<n>HetroFair uses two separate components to generate fairness-aware embeddings.<n>Our experimental results reveal that HetroFair not only alleviates unfairness and popularity bias on the item side but also achieves superior accuracy on the user side.
- Score: 1.0128808054306186
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In recent years, graph neural networks (GNNs) have become a popular tool to improve the accuracy and performance of recommender systems. Modern recommender systems are not only designed to serve end users, but also to benefit other participants, such as items and item providers. These participants may have different or conflicting goals and interests, which raises the need for fairness and popularity bias considerations. GNN-based recommendation methods also face the challenges of unfairness and popularity bias, and their normalization and aggregation processes suffer from these challenges. In this paper, we propose a fair GNN-based recommender system, called HetroFair, to improve item-side fairness. HetroFair uses two separate components to generate fairness-aware embeddings: i) Fairness-aware attention, which incorporates the dot product in the normalization process of GNNs to decrease the effect of nodes' degrees. ii) Heterophily feature weighting, to assign distinct weights to different features during the aggregation process. To evaluate the effectiveness of HetroFair, we conduct extensive experiments over six real-world datasets. Our experimental results reveal that HetroFair not only alleviates unfairness and popularity bias on the item side but also achieves superior accuracy on the user side. Our implementation is publicly available at https://github.com/NematGH/HetroFair.
Related papers
- Why Multi-Interest Fairness Matters: Hypergraph Contrastive Multi-Interest Learning for Fair Conversational Recommender System [55.39026603611269]
We propose a novel framework, Hypergraph Contrastive Multi-Interest Learning for Fair Conversational Recommender System (HyFairCRS)<n>HyFairCRS aims to promote multi-interest diversity fairness in dynamic and interactive Conversational Recommender Systems (CRSs)<n> Experiments on two CRS-based datasets show that HyFairCRS achieves a new state-of-the-art performance while effectively alleviating unfairness.
arXiv Detail & Related papers (2025-07-01T11:39:42Z) - Disentangling Popularity and Quality: An Edge Classification Approach for Fair Recommendation [1.0128808054306186]
Graph neural networks (GNNs) have proven to be an effective tool for enhancing the performance of recommender systems.
We propose a GNN-based recommendation model that disentangles popularity and quality to address this issue.
arXiv Detail & Related papers (2025-01-06T07:31:52Z) - Rethinking Fair Graph Neural Networks from Re-balancing [26.70771023446706]
We find that simple re-balancing methods can easily match or surpass existing fair GNN methods.
We propose FairGB, Fair Graph Neural Network via re-Balancing, which mitigates the unfairness of GNNs by group balancing.
arXiv Detail & Related papers (2024-07-16T11:39:27Z) - Fairpriori: Improving Biased Subgroup Discovery for Deep Neural Network Fairness [21.439820064223877]
This paper introduces Fairpriori, a novel biased subgroup discovery method.
It incorporates the frequent itemset generation algorithm to facilitate effective and efficient investigation of intersectional bias.
Fairpriori demonstrates superior effectiveness and efficiency when identifying intersectional bias.
arXiv Detail & Related papers (2024-06-25T00:15:13Z) - Marginal Debiased Network for Fair Visual Recognition [59.05212866862219]
We propose a novel marginal debiased network (MDN) to learn debiased representations.
Our MDN can achieve a remarkable performance on under-represented samples.
arXiv Detail & Related papers (2024-01-04T08:57:09Z) - Chasing Fairness in Graphs: A GNN Architecture Perspective [73.43111851492593]
We propose textsfFair textsfMessage textsfPassing (FMP) designed within a unified optimization framework for graph neural networks (GNNs)
In FMP, the aggregation is first adopted to utilize neighbors' information and then the bias mitigation step explicitly pushes demographic group node presentation centers together.
Experiments on node classification tasks demonstrate that the proposed FMP outperforms several baselines in terms of fairness and accuracy on three real-world datasets.
arXiv Detail & Related papers (2023-12-19T18:00:15Z) - Better Fair than Sorry: Adversarial Missing Data Imputation for Fair GNNs [5.655251163654288]
We propose Better Fair than Sorry, a fair missing data imputation model for protected attributes.
The key design principle behind BFtS is that imputations should approximate the worst-case scenario for fairness.
Experiments using synthetic and real datasets show that BFtS often achieves a better fairness x accuracy trade-off than existing alternatives.
arXiv Detail & Related papers (2023-11-02T20:57:44Z) - Fairness-Aware Graph Neural Networks: A Survey [53.41838868516936]
Graph Neural Networks (GNNs) have become increasingly important due to their representational power and state-of-the-art predictive performance.
GNNs suffer from fairness issues that arise as a result of the underlying graph data and the fundamental aggregation mechanism.
In this article, we examine and categorize fairness techniques for improving the fairness of GNNs.
arXiv Detail & Related papers (2023-07-08T08:09:06Z) - GFairHint: Improving Individual Fairness for Graph Neural Networks via
Fairness Hint [15.828830496326885]
algorithmic fairness in Graph Neural Networks (GNNs) has attracted significant attention.
We propose a novel method, GFairHint, which promotes individual fairness in GNNs.
GFairHint achieves the best fairness results in almost all combinations of datasets with various backbone models.
arXiv Detail & Related papers (2023-05-25T00:03:22Z) - Improving Recommendation Fairness via Data Augmentation [66.4071365614835]
Collaborative filtering based recommendation learns users' preferences from all users' historical behavior data, and has been popular to facilitate decision making.
A recommender system is considered unfair when it does not perform equally well for different user groups according to users' sensitive attributes.
In this paper, we study how to improve recommendation fairness from the data augmentation perspective.
arXiv Detail & Related papers (2023-02-13T13:11:46Z) - FedGRec: Federated Graph Recommender System with Lazy Update of Latent
Embeddings [108.77460689459247]
We propose a Federated Graph Recommender System (FedGRec) to mitigate privacy concerns.
In our system, users and the server explicitly store latent embeddings for users and items, where the latent embeddings summarize different orders of indirect user-item interactions.
We perform extensive empirical evaluations to verify the efficacy of using latent embeddings as a proxy of missing interaction graph.
arXiv Detail & Related papers (2022-10-25T01:08:20Z) - Analyzing the Effect of Sampling in GNNs on Individual Fairness [79.28449844690566]
Graph neural network (GNN) based methods have saturated the field of recommender systems.
We extend an existing method for promoting individual fairness on graphs to support mini-batch, or sub-sample based, training of a GNN.
We show that mini-batch training facilitate individual fairness promotion by allowing for local nuance to guide the process of fairness promotion in representation learning.
arXiv Detail & Related papers (2022-09-08T16:20:25Z) - FairMod: Fair Link Prediction and Recommendation via Graph Modification [7.239011273682701]
We propose FairMod to mitigate the bias learned by GNNs through modifying the input graph.
Our proposed models perform either microscopic or macroscopic edits to the input graph while training GNNs and learn node embeddings that are both accurate and fair under the context of link recommendations.
We demonstrate the effectiveness of our approach on four real world datasets and show that we can improve the recommendation fairness by several factors at negligible cost to link prediction accuracy.
arXiv Detail & Related papers (2022-01-27T15:49:33Z) - Balancing Accuracy and Fairness for Interactive Recommendation with
Reinforcement Learning [68.25805655688876]
Fairness in recommendation has attracted increasing attention due to bias and discrimination possibly caused by traditional recommenders.
We propose a reinforcement learning based framework, FairRec, to dynamically maintain a long-term balance between accuracy and fairness in IRS.
Extensive experiments validate that FairRec can improve fairness, while preserving good recommendation quality.
arXiv Detail & Related papers (2021-06-25T02:02:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.