Enhanced Doubly Robust Learning for Debiasing Post-click Conversion Rate
Estimation
- URL: http://arxiv.org/abs/2105.13623v1
- Date: Fri, 28 May 2021 06:59:49 GMT
- Title: Enhanced Doubly Robust Learning for Debiasing Post-click Conversion Rate
Estimation
- Authors: Siyuan Guo, Lixin Zou, Yiding Liu, Wenwen Ye, Suqi Cheng, Shuaiqiang
Wang, Hechang Chen, Dawei Yin, Yi Chang
- Abstract summary: Post-click conversion, as a strong signal indicating the user preference, is salutary for building recommender systems.
Currently, most existing methods utilize counterfactual learning to debias recommender systems.
We propose a novel double learning approach for the MRDR estimator, which can convert the error imputation into the general CVR estimation.
- Score: 29.27760413892272
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Post-click conversion, as a strong signal indicating the user preference, is
salutary for building recommender systems. However, accurately estimating the
post-click conversion rate (CVR) is challenging due to the selection bias,
i.e., the observed clicked events usually happen on users' preferred items.
Currently, most existing methods utilize counterfactual learning to debias
recommender systems. Among them, the doubly robust (DR) estimator has achieved
competitive performance by combining the error imputation based (EIB) estimator
and the inverse propensity score (IPS) estimator in a doubly robust way.
However, inaccurate error imputation may result in its higher variance than the
IPS estimator. Worse still, existing methods typically use simple
model-agnostic methods to estimate the imputation error, which are not
sufficient to approximate the dynamically changing model-correlated target
(i.e., the gradient direction of the prediction model). To solve these
problems, we first derive the bias and variance of the DR estimator. Based on
it, a more robust doubly robust (MRDR) estimator has been proposed to further
reduce its variance while retaining its double robustness. Moreover, we propose
a novel double learning approach for the MRDR estimator, which can convert the
error imputation into the general CVR estimation. Besides, we empirically
verify that the proposed learning scheme can further eliminate the high
variance problem of the imputation learning. To evaluate its effectiveness,
extensive experiments are conducted on a semi-synthetic dataset and two
real-world datasets. The results demonstrate the superiority of the proposed
approach over the state-of-the-art methods. The code is available at
https://github.com/guosyjlu/MRDR-DL.
Related papers
- Debiased Recommendation with Noisy Feedback [41.38490962524047]
We study intersectional threats to the unbiased learning of the prediction model from data MNAR and OME in the collected data.
First, we design OME-EIB, OME-IPS, and OME-DR estimators, which largely extend the existing estimators to combat OME in real-world recommendation scenarios.
arXiv Detail & Related papers (2024-06-24T23:42:18Z) - A Generalized Doubly Robust Learning Framework for Debiasing Post-Click
Conversion Rate Prediction [23.340584290411208]
Post-click conversion rate (CVR) prediction is an essential task for discovering user interests and increasing platform revenues.
Currently, doubly robust (DR) learning approaches achieve the state-of-the-art performance for debiasing CVR prediction.
We propose two new DR methods, namely DR-BIAS and DR-MSE, which control the bias of DR loss and balance the bias and variance flexibly.
arXiv Detail & Related papers (2022-11-12T15:09:23Z) - Rethinking Missing Data: Aleatoric Uncertainty-Aware Recommendation [59.500347564280204]
We propose a new Aleatoric Uncertainty-aware Recommendation (AUR) framework.
AUR consists of a new uncertainty estimator along with a normal recommender model.
As the chance of mislabeling reflects the potential of a pair, AUR makes recommendations according to the uncertainty.
arXiv Detail & Related papers (2022-09-22T04:32:51Z) - Multiple Robust Learning for Recommendation [13.06593469196849]
In recommender systems, a common problem is the presence of various biases in the collected data.
We propose a multiple robust (MR) estimator that can take the advantage of multiple candidate imputation and propensity models to achieve unbiasedness.
arXiv Detail & Related papers (2022-07-09T13:15:56Z) - Debiasing Learning for Membership Inference Attacks Against Recommender
Systems [79.48353547307887]
Learned recommender systems may inadvertently leak information about their training data, leading to privacy violations.
We investigate privacy threats faced by recommender systems through the lens of membership inference.
We propose a Debiasing Learning for Membership Inference Attacks against recommender systems (DL-MIA) framework that has four main components.
arXiv Detail & Related papers (2022-06-24T17:57:34Z) - StableDR: Stabilized Doubly Robust Learning for Recommendation on Data
Missing Not at Random [16.700598755439685]
We show that the doubly robust (DR) methods are unstable and have unbounded bias, variance, and generalization bounds to extremely small propensities.
We propose a doubly robust (StableDR) learning approach with a weaker reliance on extrapolation.
In addition, we propose a novel learning approach for StableDR that updates the imputation, propensity, and prediction models cyclically.
arXiv Detail & Related papers (2022-05-10T07:04:53Z) - Cross Pairwise Ranking for Unbiased Item Recommendation [57.71258289870123]
We develop a new learning paradigm named Cross Pairwise Ranking (CPR)
CPR achieves unbiased recommendation without knowing the exposure mechanism.
We prove in theory that this way offsets the influence of user/item propensity on the learning.
arXiv Detail & Related papers (2022-04-26T09:20:27Z) - Doubly Robust Collaborative Targeted Learning for Recommendation on Data
Missing Not at Random [6.563595953273317]
In recommender systems, the feedback data received is always missing not at random (MNAR)
We propose bf DR-TMLE that effectively captures the merits of both error imputation-based (EIB) and doubly robust (DR) methods.
We also propose a novel RCT-free collaborative targeted learning algorithm for DR-TMLE, called bf DR-TMLE-TL
arXiv Detail & Related papers (2022-03-19T06:48:50Z) - Doubly Robust Distributionally Robust Off-Policy Evaluation and Learning [59.02006924867438]
Off-policy evaluation and learning (OPE/L) use offline observational data to make better decisions.
Recent work proposed distributionally robust OPE/L (DROPE/L) to remedy this, but the proposal relies on inverse-propensity weighting.
We propose the first DR algorithms for DROPE/L with KL-divergence uncertainty sets.
arXiv Detail & Related papers (2022-02-19T20:00:44Z) - Learning to Estimate Without Bias [57.82628598276623]
Gauss theorem states that the weighted least squares estimator is a linear minimum variance unbiased estimation (MVUE) in linear models.
In this paper, we take a first step towards extending this result to non linear settings via deep learning with bias constraints.
A second motivation to BCE is in applications where multiple estimates of the same unknown are averaged for improved performance.
arXiv Detail & Related papers (2021-10-24T10:23:51Z) - Machine learning for causal inference: on the use of cross-fit
estimators [77.34726150561087]
Doubly-robust cross-fit estimators have been proposed to yield better statistical properties.
We conducted a simulation study to assess the performance of several estimators for the average causal effect (ACE)
When used with machine learning, the doubly-robust cross-fit estimators substantially outperformed all of the other estimators in terms of bias, variance, and confidence interval coverage.
arXiv Detail & Related papers (2020-04-21T23:09:55Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.