Causality-aware Graph Aggregation Weight Estimator for Popularity Debiasing in Top-K Recommendation
- URL: http://arxiv.org/abs/2510.04502v1
- Date: Mon, 06 Oct 2025 05:33:37 GMT
- Title: Causality-aware Graph Aggregation Weight Estimator for Popularity Debiasing in Top-K Recommendation
- Authors: Yue Que, Yingyi Zhang, Xiangyu Zhao, Chen Ma,
- Abstract summary: Graph-based recommender systems leverage neighborhood aggregation to generate node representations, which is sensitive to popularity bias.<n>Existing graph-based debiasing solutions refine the aggregation process with attempts such as edge reconstruction or weight adjustment.<n>We propose a novel approach to mitigate popularity bias through rational modeling of the graph aggregation process.
- Score: 20.03303865662207
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Graph-based recommender systems leverage neighborhood aggregation to generate node representations, which is highly sensitive to popularity bias, resulting in an echo effect during information propagation. Existing graph-based debiasing solutions refine the aggregation process with attempts such as edge reconstruction or weight adjustment. However, these methods remain inadequate in fully alleviating popularity bias. Specifically, this is because 1) they provide no insights into graph aggregation rationality, thus lacking an optimality guarantee; 2) they fail to well balance the training and debiasing process, which undermines the effectiveness. In this paper, we propose a novel approach to mitigate popularity bias through rational modeling of the graph aggregation process. We reveal that graph aggregation is a special form of backdoor adjustment in causal inference, where the aggregation weight corresponds to the historical interaction likelihood distribution. Based on this insight, we devise an encoder-decoder architecture, namely Causality-aware Graph Aggregation Weight Estimator for Debiasing (CAGED), to approximate the unbiased aggregation weight by optimizing the evidence lower bound of the interaction likelihood. In order to enhance the debiasing effectiveness during early training stages, we further design a momentum update strategy that incrementally refines the aggregation weight matrix. Extensive experiments on three datasets demonstrate that CAGED outperforms existing graph-based debiasing methods. Our implementation is available at https://github.com/QueYork/CAGED.
Related papers
- Graph Negative Feedback Bias Correction Framework for Adaptive Heterophily Modeling [15.157616444432563]
Graph Neural Networks (GNNs) have emerged as a powerful framework for processing graph-structured data.<n>GNNs are inherently limited by the homophily assumption, leading to degradation in performance on heterophilic graphs.<n>We propose Graph Negative Feedback Bias Correction (GNFBC), a simple yet effective framework that is independent of any specific aggregation strategy.
arXiv Detail & Related papers (2026-03-04T02:34:48Z) - Chasing Fairness in Graphs: A GNN Architecture Perspective [73.43111851492593]
We propose textsfFair textsfMessage textsfPassing (FMP) designed within a unified optimization framework for graph neural networks (GNNs)
In FMP, the aggregation is first adopted to utilize neighbors' information and then the bias mitigation step explicitly pushes demographic group node presentation centers together.
Experiments on node classification tasks demonstrate that the proposed FMP outperforms several baselines in terms of fairness and accuracy on three real-world datasets.
arXiv Detail & Related papers (2023-12-19T18:00:15Z) - Fairness-aware Optimal Graph Filter Design [25.145533328758614]
Graphs are mathematical tools that can be used to represent complex real-world interconnected systems.
Machine learning (ML) over graphs has attracted significant attention recently.
We take a fresh look at the problem of bias mitigation in graph-based learning by borrowing insights from graph signal processing.
arXiv Detail & Related papers (2023-10-22T22:40:40Z) - Efficient Bi-Level Optimization for Recommendation Denoising [31.968068788022403]
implicit feedback possesses a high degree of noise, which significantly undermines recommendation quality.
We model recommendation denoising as a bi-level optimization problem.
The inner optimization aims to derive an effective model for the recommendation, as well as guiding the weight determination.
We employ a weight generator to avoid the storage of weights and a one-step gradient-matching-based loss to significantly reduce computational time.
arXiv Detail & Related papers (2022-10-19T06:36:21Z) - Kernel-Whitening: Overcome Dataset Bias with Isotropic Sentence
Embedding [51.48582649050054]
We propose a representation normalization method which aims at disentangling the correlations between features of encoded sentences.
We also propose Kernel-Whitening, a Nystrom kernel approximation method to achieve more thorough debiasing on nonlinear spurious correlations.
Experiments show that Kernel-Whitening significantly improves the performance of BERT on out-of-distribution datasets while maintaining in-distribution accuracy.
arXiv Detail & Related papers (2022-10-14T05:56:38Z) - Optimal Propagation for Graph Neural Networks [51.08426265813481]
We propose a bi-level optimization approach for learning the optimal graph structure.
We also explore a low-rank approximation model for further reducing the time complexity.
arXiv Detail & Related papers (2022-05-06T03:37:00Z) - Causal Incremental Graph Convolution for Recommender System Retraining [89.25922726558875]
Real-world recommender system needs to be regularly retrained to keep with the new data.
In this work, we consider how to efficiently retrain graph convolution network (GCN) based recommender models.
arXiv Detail & Related papers (2021-08-16T04:20:09Z) - Training Robust Graph Neural Networks with Topology Adaptive Edge
Dropping [116.26579152942162]
Graph neural networks (GNNs) are processing architectures that exploit graph structural information to model representations from network data.
Despite their success, GNNs suffer from sub-optimal generalization performance given limited training data.
This paper proposes Topology Adaptive Edge Dropping to improve generalization performance and learn robust GNN models.
arXiv Detail & Related papers (2021-06-05T13:20:36Z) - Graph Classification by Mixture of Diverse Experts [67.33716357951235]
We present GraphDIVE, a framework leveraging mixture of diverse experts for imbalanced graph classification.
With a divide-and-conquer principle, GraphDIVE employs a gating network to partition an imbalanced graph dataset into several subsets.
Experiments on real-world imbalanced graph datasets demonstrate the effectiveness of GraphDIVE.
arXiv Detail & Related papers (2021-03-29T14:03:03Z) - Heuristic Semi-Supervised Learning for Graph Generation Inspired by
Electoral College [80.67842220664231]
We propose a novel pre-processing technique, namely ELectoral COllege (ELCO), which automatically expands new nodes and edges to refine the label similarity within a dense subgraph.
In all setups tested, our method boosts the average score of base models by a large margin of 4.7 points, as well as consistently outperforms the state-of-the-art.
arXiv Detail & Related papers (2020-06-10T14:48:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.