Challenges and approaches to privacy preserving post-click conversion
prediction
- URL: http://arxiv.org/abs/2201.12666v1
- Date: Sat, 29 Jan 2022 21:36:01 GMT
- Title: Challenges and approaches to privacy preserving post-click conversion
prediction
- Authors: Conor O'Brien, Arvind Thiagarajan, Sourav Das, Rafael Barreto, Chetan
Verma, Tim Hsu, James Neufield, Jonathan J Hunt
- Abstract summary: We provide an overview of the challenges and constraints when learning conversion models in this setting.
We introduce a novel approach for training these models that makes use of post-ranking signals.
We show using offline experiments on real world data that it outperforms a model relying on opt-in data alone.
- Score: 3.4071263815701336
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Online advertising has typically been more personalized than offline
advertising, through the use of machine learning models and real-time auctions
for ad targeting. One specific task, predicting the likelihood of conversion
(i.e.\ the probability a user will purchase the advertised product), is crucial
to the advertising ecosystem for both targeting and pricing ads. Currently,
these models are often trained by observing individual user behavior, but,
increasingly, regulatory and technical constraints are requiring
privacy-preserving approaches. For example, major platforms are moving to
restrict tracking individual user events across multiple applications, and
governments around the world have shown steadily more interest in regulating
the use of personal data. Instead of receiving data about individual user
behavior, advertisers may receive privacy-preserving feedback, such as the
number of installs of an advertised app that resulted from a group of users. In
this paper we outline the recent privacy-related changes in the online
advertising ecosystem from a machine learning perspective. We provide an
overview of the challenges and constraints when learning conversion models in
this setting. We introduce a novel approach for training these models that
makes use of post-ranking signals. We show using offline experiments on real
world data that it outperforms a model relying on opt-in data alone, and
significantly reduces model degradation when no individual labels are
available. Finally, we discuss future directions for research in this evolving
area.
Related papers
- Privacy Backdoors: Enhancing Membership Inference through Poisoning Pre-trained Models [112.48136829374741]
In this paper, we unveil a new vulnerability: the privacy backdoor attack.
When a victim fine-tunes a backdoored model, their training data will be leaked at a significantly higher rate than if they had fine-tuned a typical model.
Our findings highlight a critical privacy concern within the machine learning community and call for a reevaluation of safety protocols in the use of open-source pre-trained models.
arXiv Detail & Related papers (2024-04-01T16:50:54Z) - Protecting User Privacy in Online Settings via Supervised Learning [69.38374877559423]
We design an intelligent approach to online privacy protection that leverages supervised learning.
By detecting and blocking data collection that might infringe on a user's privacy, we can restore a degree of digital privacy to the user.
arXiv Detail & Related papers (2023-04-06T05:20:16Z) - Client-specific Property Inference against Secure Aggregation in
Federated Learning [52.8564467292226]
Federated learning has become a widely used paradigm for collaboratively training a common model among different participants.
Many attacks have shown that it is still possible to infer sensitive information such as membership, property, or outright reconstruction of participant data.
We show that simple linear models can effectively capture client-specific properties only from the aggregated model updates.
arXiv Detail & Related papers (2023-03-07T14:11:01Z) - Towards a User Privacy-Aware Mobile Gaming App Installation Prediction
Model [0.8602553195689513]
We investigate the process of predicting a mobile gaming app installation from the point of view of a demand-side platform.
We explore the trade-off between privacy preservation and model performance.
We conclude that privacy-aware models might still preserve significant capabilities.
arXiv Detail & Related papers (2023-02-07T09:14:59Z) - Position: Considerations for Differentially Private Learning with Large-Scale Public Pretraining [75.25943383604266]
We question whether the use of large Web-scraped datasets should be viewed as differential-privacy-preserving.
We caution that publicizing these models pretrained on Web data as "private" could lead to harm and erode the public's trust in differential privacy as a meaningful definition of privacy.
We conclude by discussing potential paths forward for the field of private learning, as public pretraining becomes more popular and powerful.
arXiv Detail & Related papers (2022-12-13T10:41:12Z) - Canary in a Coalmine: Better Membership Inference with Ensembled
Adversarial Queries [53.222218035435006]
We use adversarial tools to optimize for queries that are discriminative and diverse.
Our improvements achieve significantly more accurate membership inference than existing methods.
arXiv Detail & Related papers (2022-10-19T17:46:50Z) - Lessons from the AdKDD'21 Privacy-Preserving ML Challenge [57.365745458033075]
A prominent proposal at W3C only allows sharing advertising signals through aggregated, differentially private reports of past displays.
To study this proposal extensively, an open Privacy-Preserving Machine Learning Challenge took place at AdKDD'21.
A key finding is that learning models on large, aggregated data in the presence of a small set of unaggregated data points can be surprisingly efficient and cheap.
arXiv Detail & Related papers (2022-01-31T11:09:59Z) - Improving Fairness and Privacy in Selection Problems [21.293367386282902]
We study the possibility of using a differentially private exponential mechanism as a post-processing step to improve both fairness and privacy of supervised learning models.
We show that the exponential mechanism can improve both privacy and fairness, with a slight decrease in accuracy compared to the model without post-processing.
arXiv Detail & Related papers (2020-12-07T15:55:28Z) - Privacy Enhancing Machine Learning via Removal of Unwanted Dependencies [21.97951347784442]
This paper studies new variants of supervised and adversarial learning methods, which remove the sensitive information in the data before they are sent out for a particular application.
The explored methods optimize privacy preserving feature mappings and predictive models simultaneously in an end-to-end fashion.
Experimental results on mobile sensing and face datasets demonstrate that our models can successfully maintain the utility performances of predictive models while causing sensitive predictions to perform poorly.
arXiv Detail & Related papers (2020-07-30T19:55:10Z) - Fairness-Aware Online Personalization [16.320648868892526]
We present a study of fairness in online personalization settings involving the ranking of individuals.
We first demonstrate that online personalization can cause the model to learn to act in an unfair manner if the user is biased in his/her responses.
We then formulate the problem of learning personalized models under fairness constraints and present a regularization based approach for mitigating biases in machine learning.
arXiv Detail & Related papers (2020-07-30T07:16:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.