Conversion Rate Prediction via Meta Learning in Small-Scale
Recommendation Scenarios
- URL: http://arxiv.org/abs/2112.13753v1
- Date: Mon, 27 Dec 2021 16:05:42 GMT
- Title: Conversion Rate Prediction via Meta Learning in Small-Scale
Recommendation Scenarios
- Authors: Xiaofeng Pan, Ming Li, Jing Zhang, Keren Yu, Luping Wang, Hong Wen,
Chengjun Mao and Bo Cao
- Abstract summary: We propose a novel CVR method named MetaCVR from a perspective of meta learning to address the Data Distribution Fluctuation (DDF) issue.
To the best of our knowledge, this is the first study of CVR prediction targeting the DDF issue in small-scale recommendation scenarios.
- Score: 17.02759665047561
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Different from large-scale platforms such as Taobao and Amazon, developing
CVR models in small-scale recommendation scenarios is more challenging due to
the severe Data Distribution Fluctuation (DDF) issue. DDF prevents existing CVR
models from being effective since 1) several months of data are needed to train
CVR models sufficiently in small scenarios, leading to considerable
distribution discrepancy between training and online serving; and 2) e-commerce
promotions have much more significant impacts on small scenarios, leading to
distribution uncertainty of the upcoming time period. In this work, we propose
a novel CVR method named MetaCVR from a perspective of meta learning to address
the DDF issue. Firstly, a base CVR model which consists of a Feature
Representation Network (FRN) and output layers is elaborately designed and
trained sufficiently with samples across months. Then we treat time periods
with different data distributions as different occasions and obtain positive
and negative prototypes for each occasion using the corresponding samples and
the pre-trained FRN. Subsequently, a Distance Metric Network (DMN) is devised
to calculate the distance metrics between each sample and all prototypes to
facilitate mitigating the distribution uncertainty. At last, we develop an
Ensemble Prediction Network (EPN) which incorporates the output of FRN and DMN
to make the final CVR prediction. In this stage, we freeze the FRN and train
the DMN and EPN with samples from recent time period, therefore effectively
easing the distribution discrepancy. To the best of our knowledge, this is the
first study of CVR prediction targeting the DDF issue in small-scale
recommendation scenarios. Experimental results on real-world datasets validate
the superiority of our MetaCVR and online A/B test also shows our model
achieves impressive gains of 11.92% on PCVR and 8.64% on GMV.
Related papers
- DOTA: Distributional Test-Time Adaptation of Vision-Language Models [52.98590762456236]
Training-free test-time dynamic adapter (TDA) is a promising approach to address this issue.
We propose a simple yet effective method for DistributiOnal Test-time Adaptation (Dota)
Dota continually estimates the distributions of test samples, allowing the model to continually adapt to the deployment environment.
arXiv Detail & Related papers (2024-09-28T15:03:28Z) - Variational Mode Decomposition and Linear Embeddings are What You Need For Time-Series Forecasting [2.1233286062376497]
Variational Mode Decomposition (VMD) has emerged as a promising technique to mitigate volatility by decomposing data into distinct modes.
In this study, we integrate VMD with linear models to develop a robust forecasting framework.
arXiv Detail & Related papers (2024-08-28T20:22:09Z) - Probabilistic Contrastive Learning for Long-Tailed Visual Recognition [78.70453964041718]
Longtailed distributions frequently emerge in real-world data, where a large number of minority categories contain a limited number of samples.
Recent investigations have revealed that supervised contrastive learning exhibits promising potential in alleviating the data imbalance.
We propose a novel probabilistic contrastive (ProCo) learning algorithm that estimates the data distribution of the samples from each class in the feature space.
arXiv Detail & Related papers (2024-03-11T13:44:49Z) - Contrastive Learning for Conversion Rate Prediction [6.607531486024888]
We propose Contrastive Learning for CVR prediction (CL4CVR) framework.
It associates the supervised CVR prediction task with a contrastive learning task, which can learn better data representations.
Experimental results on two real-world conversion datasets demonstrate the superior performance of CL4CVR.
arXiv Detail & Related papers (2023-07-12T07:42:52Z) - Conservative Prediction via Data-Driven Confidence Minimization [70.93946578046003]
In safety-critical applications of machine learning, it is often desirable for a model to be conservative.
We propose the Data-Driven Confidence Minimization framework, which minimizes confidence on an uncertainty dataset.
arXiv Detail & Related papers (2023-06-08T07:05:36Z) - Leveraging Unlabeled Data to Predict Out-of-Distribution Performance [63.740181251997306]
Real-world machine learning deployments are characterized by mismatches between the source (training) and target (test) distributions.
In this work, we investigate methods for predicting the target domain accuracy using only labeled source data and unlabeled target data.
We propose Average Thresholded Confidence (ATC), a practical method that learns a threshold on the model's confidence, predicting accuracy as the fraction of unlabeled examples.
arXiv Detail & Related papers (2022-01-11T23:01:12Z) - Empirical Models for Multidimensional Regression of Fission Systems [0.0]
We develop guidelines for developing empirical models for multidimensional regression of neutron transport.
An assessment of the accuracy and precision finds that the SVR, followed closely by ANN, performs the best.
arXiv Detail & Related papers (2021-05-30T22:53:39Z) - Delayed Feedback Modeling for the Entire Space Conversion Rate
Prediction [15.579755993971657]
Estimating post-click conversion rate (CVR) accurately is crucial in E-commerce.
We propose a novel neural network framework ESDF to tackle the above three challenges simultaneously.
arXiv Detail & Related papers (2020-11-24T01:14:03Z) - Robust Validation: Confident Predictions Even When Distributions Shift [19.327409270934474]
We describe procedures for robust predictive inference, where a model provides uncertainty estimates on its predictions rather than point predictions.
We present a method that produces prediction sets (almost exactly) giving the right coverage level for any test distribution in an $f$-divergence ball around the training population.
An essential component of our methodology is to estimate the amount of expected future data shift and build robustness to it.
arXiv Detail & Related papers (2020-08-10T17:09:16Z) - Diversity inducing Information Bottleneck in Model Ensembles [73.80615604822435]
In this paper, we target the problem of generating effective ensembles of neural networks by encouraging diversity in prediction.
We explicitly optimize a diversity inducing adversarial loss for learning latent variables and thereby obtain diversity in the output predictions necessary for modeling multi-modal data.
Compared to the most competitive baselines, we show significant improvements in classification accuracy, under a shift in the data distribution.
arXiv Detail & Related papers (2020-03-10T03:10:41Z) - Meta-Learned Confidence for Few-shot Learning [60.6086305523402]
A popular transductive inference technique for few-shot metric-based approaches, is to update the prototype of each class with the mean of the most confident query examples.
We propose to meta-learn the confidence for each query sample, to assign optimal weights to unlabeled queries.
We validate our few-shot learning model with meta-learned confidence on four benchmark datasets.
arXiv Detail & Related papers (2020-02-27T10:22:17Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.