Unbiased Filtering Of Accidental Clicks in Verizon Media Native
Advertising
- URL: http://arxiv.org/abs/2312.05017v1
- Date: Fri, 8 Dec 2023 12:54:30 GMT
- Title: Unbiased Filtering Of Accidental Clicks in Verizon Media Native
Advertising
- Authors: Yohay Kaplan and Naama Krasne and Alex Shtoff and Oren Somekh
- Abstract summary: We focus on the challenge of predicting click-through rates (CTR) when we are aware that some of the clicks have short dwell-time.
An accidental click implies little affinity between the user and the ad, so predicting that similar users will click on the ad is inaccurate.
We present a new approach where the positive weight of the accidental clicks is distributed among all of the negative events (skips), based on their likelihood of causing accidental clicks.
- Score: 1.6717433307723157
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Verizon Media (VZM) native advertising is one of VZM largest and fastest
growing businesses, reaching a run-rate of several hundred million USDs in the
past year. Driving the VZM native models that are used to predict event
probabilities, such as click and conversion probabilities, is OFFSET - a
feature enhanced collaborative-filtering based event-prediction algorithm. In
this work we focus on the challenge of predicting click-through rates (CTR)
when we are aware that some of the clicks have short dwell-time and are defined
as accidental clicks. An accidental click implies little affinity between the
user and the ad, so predicting that similar users will click on the ad is
inaccurate. Therefore, it may be beneficial to remove clicks with dwell-time
lower than a predefined threshold from the training set. However, we cannot
ignore these positive events, as filtering these will cause the model to under
predict. Previous approaches have tried to apply filtering and then adding
corrective biases to the CTR predictions, but did not yield revenue lifts and
therefore were not adopted. In this work, we present a new approach where the
positive weight of the accidental clicks is distributed among all of the
negative events (skips), based on their likelihood of causing accidental
clicks, as predicted by an auxiliary model. These likelihoods are taken as the
correct labels of the negative events, shifting our training from using only
binary labels and adopting a binary cross-entropy loss function in our training
process. After showing offline performance improvements, the modified model was
tested online serving VZM native users, and provided 1.18% revenue lift over
the production model which is agnostic to accidental clicks.
Related papers
- Unbiased Learning to Rank with Query-Level Click Propensity Estimation: Beyond Pointwise Observation and Relevance [74.43264459255121]
In real-world scenarios, users often click only one or two results after examining multiple relevant options.
We propose a query-level click propensity model to capture the probability that users will click on different result lists.
Our method introduces a Dual Inverse Propensity Weighting mechanism to address both relevance saturation and position bias.
arXiv Detail & Related papers (2025-02-17T03:55:51Z) - Enhancing Zero-Shot Vision Models by Label-Free Prompt Distribution Learning and Bias Correcting [55.361337202198925]
Vision-language models, such as CLIP, have shown impressive generalization capacities when using appropriate text descriptions.
We propose a label-Free prompt distribution learning and bias correction framework, dubbed as **Frolic**, which boosts zero-shot performance without the need for labeled data.
arXiv Detail & Related papers (2024-10-25T04:00:45Z) - Improved Online Learning Algorithms for CTR Prediction in Ad Auctions [8.2536631346421]
We investigate the online learning problem of revenue in ad auctions.
We focus on two models of the advertisers' strategic behaviors.
We develop an online mechanism based on upper-confidence bounds that achieves a tight $O(sqrtT)$ regret.
arXiv Detail & Related papers (2024-02-29T14:10:26Z) - Improving conversion rate prediction via self-supervised pre-training in
online advertising [2.447795279790662]
Key challenge in training models that predict conversions-given-clicks comes from data sparsity.
We use the well-known idea of self-supervised pre-training, and use an auxiliary auto-encoder model trained on all conversion events.
We show improvements both offline, during training, and in an online A/B test.
arXiv Detail & Related papers (2024-01-25T08:44:22Z) - Soft Frequency Capping for Improved Ad Click Prediction in Yahoo Gemini
Native [1.9315883475944244]
Yahoo's native advertising (also known as Gemini native) serves billions of ad impressions daily, reaching a yearly run-rate of many hundred of millions USD.
Offset is a one-pass algorithm that updates its model for every new batch of logged data using a gradient descent (SGD) based approach.
We propose a soft frequency capping (SFC) approach, where the frequency feature is incorporated into the OFFSET model as a user-ad feature and its weight vector is learned via logistic regression.
arXiv Detail & Related papers (2023-12-08T14:12:49Z) - Click-Conversion Multi-Task Model with Position Bias Mitigation for
Sponsored Search in eCommerce [51.211924408864355]
We propose two position-bias-free prediction models: Position-Aware Click-Conversion (PACC) and PACC via Position Embedding (PACC-PE)
Experiments on the E-commerce sponsored product search dataset show that our proposed models have better ranking effectiveness and can greatly alleviate position bias in both CTR and CVR prediction.
arXiv Detail & Related papers (2023-07-29T19:41:16Z) - An Offline Metric for the Debiasedness of Click Models [52.25681483524383]
Click models are a common method for extracting information from user clicks.
Recent work shows that the current evaluation practices in the community fail to guarantee that a well-performing click model generalizes well to downstream tasks.
We introduce the concept of debiasedness in click modeling and derive a metric for measuring it.
arXiv Detail & Related papers (2023-04-19T10:59:34Z) - Meta-Wrapper: Differentiable Wrapping Operator for User Interest
Selection in CTR Prediction [97.99938802797377]
Click-through rate (CTR) prediction, whose goal is to predict the probability of the user to click on an item, has become increasingly significant in recommender systems.
Recent deep learning models with the ability to automatically extract the user interest from his/her behaviors have achieved great success.
We propose a novel approach under the framework of the wrapper method, which is named Meta-Wrapper.
arXiv Detail & Related papers (2022-06-28T03:28:15Z) - Scalar is Not Enough: Vectorization-based Unbiased Learning to Rank [29.934700345584726]
Unbiased learning to rank aims to train an unbiased ranking model from biased user click logs.
Most of the current ULTR methods are based on the examination hypothesis (EH), which assumes that the click probability can be factorized into two scalar functions.
We propose a vector-based EH and formulate the click probability as a dot product of two vector functions.
arXiv Detail & Related papers (2022-06-03T17:23:25Z) - Almost Tight L0-norm Certified Robustness of Top-k Predictions against
Adversarial Perturbations [78.23408201652984]
Top-k predictions are used in many real-world applications such as machine learning as a service, recommender systems, and web searches.
Our work is based on randomized smoothing, which builds a provably robust classifier via randomizing an input.
For instance, our method can build a classifier that achieves a certified top-3 accuracy of 69.2% on ImageNet when an attacker can arbitrarily perturb 5 pixels of a testing image.
arXiv Detail & Related papers (2020-11-15T21:34:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.