Unbiased Filtering Of Accidental Clicks in Verizon Media Native
Advertising
- URL: http://arxiv.org/abs/2312.05017v1
- Date: Fri, 8 Dec 2023 12:54:30 GMT
- Title: Unbiased Filtering Of Accidental Clicks in Verizon Media Native
Advertising
- Authors: Yohay Kaplan and Naama Krasne and Alex Shtoff and Oren Somekh
- Abstract summary: We focus on the challenge of predicting click-through rates (CTR) when we are aware that some of the clicks have short dwell-time.
An accidental click implies little affinity between the user and the ad, so predicting that similar users will click on the ad is inaccurate.
We present a new approach where the positive weight of the accidental clicks is distributed among all of the negative events (skips), based on their likelihood of causing accidental clicks.
- Score: 1.6717433307723157
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Verizon Media (VZM) native advertising is one of VZM largest and fastest
growing businesses, reaching a run-rate of several hundred million USDs in the
past year. Driving the VZM native models that are used to predict event
probabilities, such as click and conversion probabilities, is OFFSET - a
feature enhanced collaborative-filtering based event-prediction algorithm. In
this work we focus on the challenge of predicting click-through rates (CTR)
when we are aware that some of the clicks have short dwell-time and are defined
as accidental clicks. An accidental click implies little affinity between the
user and the ad, so predicting that similar users will click on the ad is
inaccurate. Therefore, it may be beneficial to remove clicks with dwell-time
lower than a predefined threshold from the training set. However, we cannot
ignore these positive events, as filtering these will cause the model to under
predict. Previous approaches have tried to apply filtering and then adding
corrective biases to the CTR predictions, but did not yield revenue lifts and
therefore were not adopted. In this work, we present a new approach where the
positive weight of the accidental clicks is distributed among all of the
negative events (skips), based on their likelihood of causing accidental
clicks, as predicted by an auxiliary model. These likelihoods are taken as the
correct labels of the negative events, shifting our training from using only
binary labels and adopting a binary cross-entropy loss function in our training
process. After showing offline performance improvements, the modified model was
tested online serving VZM native users, and provided 1.18% revenue lift over
the production model which is agnostic to accidental clicks.
Related papers
- Enhancing Zero-Shot Vision Models by Label-Free Prompt Distribution Learning and Bias Correcting [55.361337202198925]
Vision-language models, such as CLIP, have shown impressive generalization capacities when using appropriate text descriptions.
We propose a label-Free prompt distribution learning and bias correction framework, dubbed as **Frolic**, which boosts zero-shot performance without the need for labeled data.
arXiv Detail & Related papers (2024-10-25T04:00:45Z) - Improved Online Learning Algorithms for CTR Prediction in Ad Auctions [8.2536631346421]
We investigate the online learning problem of revenue in ad auctions.
We focus on two models of the advertisers' strategic behaviors.
We develop an online mechanism based on upper-confidence bounds that achieves a tight $O(sqrtT)$ regret.
arXiv Detail & Related papers (2024-02-29T14:10:26Z) - Improving conversion rate prediction via self-supervised pre-training in
online advertising [2.447795279790662]
Key challenge in training models that predict conversions-given-clicks comes from data sparsity.
We use the well-known idea of self-supervised pre-training, and use an auxiliary auto-encoder model trained on all conversion events.
We show improvements both offline, during training, and in an online A/B test.
arXiv Detail & Related papers (2024-01-25T08:44:22Z) - Soft Frequency Capping for Improved Ad Click Prediction in Yahoo Gemini
Native [1.9315883475944244]
Yahoo's native advertising (also known as Gemini native) serves billions of ad impressions daily, reaching a yearly run-rate of many hundred of millions USD.
Offset is a one-pass algorithm that updates its model for every new batch of logged data using a gradient descent (SGD) based approach.
We propose a soft frequency capping (SFC) approach, where the frequency feature is incorporated into the OFFSET model as a user-ad feature and its weight vector is learned via logistic regression.
arXiv Detail & Related papers (2023-12-08T14:12:49Z) - Click-Conversion Multi-Task Model with Position Bias Mitigation for
Sponsored Search in eCommerce [51.211924408864355]
We propose two position-bias-free prediction models: Position-Aware Click-Conversion (PACC) and PACC via Position Embedding (PACC-PE)
Experiments on the E-commerce sponsored product search dataset show that our proposed models have better ranking effectiveness and can greatly alleviate position bias in both CTR and CVR prediction.
arXiv Detail & Related papers (2023-07-29T19:41:16Z) - Rec4Ad: A Free Lunch to Mitigate Sample Selection Bias for Ads CTR
Prediction in Taobao [24.43583745735832]
We propose to leverage recommendations samples as a free lunch to mitigate sample selection bias for ads CTR model (Rec4Ad)
Rec4Ad achieves substantial gains in key business metrics, with a lift of up to +6.6% CTR and +2.9% RPM.
arXiv Detail & Related papers (2023-06-06T09:22:52Z) - ASPEST: Bridging the Gap Between Active Learning and Selective
Prediction [56.001808843574395]
Selective prediction aims to learn a reliable model that abstains from making predictions when uncertain.
Active learning aims to lower the overall labeling effort, and hence human dependence, by querying the most informative examples.
In this work, we introduce a new learning paradigm, active selective prediction, which aims to query more informative samples from the shifted target domain.
arXiv Detail & Related papers (2023-04-07T23:51:07Z) - Meta-Wrapper: Differentiable Wrapping Operator for User Interest
Selection in CTR Prediction [97.99938802797377]
Click-through rate (CTR) prediction, whose goal is to predict the probability of the user to click on an item, has become increasingly significant in recommender systems.
Recent deep learning models with the ability to automatically extract the user interest from his/her behaviors have achieved great success.
We propose a novel approach under the framework of the wrapper method, which is named Meta-Wrapper.
arXiv Detail & Related papers (2022-06-28T03:28:15Z) - Scalar is Not Enough: Vectorization-based Unbiased Learning to Rank [29.934700345584726]
Unbiased learning to rank aims to train an unbiased ranking model from biased user click logs.
Most of the current ULTR methods are based on the examination hypothesis (EH), which assumes that the click probability can be factorized into two scalar functions.
We propose a vector-based EH and formulate the click probability as a dot product of two vector functions.
arXiv Detail & Related papers (2022-06-03T17:23:25Z) - Taming Overconfident Prediction on Unlabeled Data from Hindsight [50.9088560433925]
Minimizing prediction uncertainty on unlabeled data is a key factor to achieve good performance in semi-supervised learning.
This paper proposes a dual mechanism, named ADaptive Sharpening (ADS), which first applies a soft-threshold to adaptively mask out determinate and negligible predictions.
ADS significantly improves the state-of-the-art SSL methods by making it a plug-in.
arXiv Detail & Related papers (2021-12-15T15:17:02Z) - Almost Tight L0-norm Certified Robustness of Top-k Predictions against
Adversarial Perturbations [78.23408201652984]
Top-k predictions are used in many real-world applications such as machine learning as a service, recommender systems, and web searches.
Our work is based on randomized smoothing, which builds a provably robust classifier via randomizing an input.
For instance, our method can build a classifier that achieves a certified top-3 accuracy of 69.2% on ImageNet when an attacker can arbitrarily perturb 5 pixels of a testing image.
arXiv Detail & Related papers (2020-11-15T21:34:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.