Contrastive Multi-view Framework for Customer Lifetime Value Prediction
- URL: http://arxiv.org/abs/2306.14400v1
- Date: Mon, 26 Jun 2023 03:23:53 GMT
- Title: Contrastive Multi-view Framework for Customer Lifetime Value Prediction
- Authors: Chuhan Wu, Jingjie Li, Qinglin Jia, Hong Zhu, Yuan Fang and Ruiming
Tang
- Abstract summary: Many existing LTV prediction methods directly train a single-view LTV predictor on consumption samples.
We propose a contrastive multi-view framework for LTV prediction, which is a plug-and-play solution compatible with various backbone models.
We conduct extensive experiments on a real-world game LTV prediction dataset and the results validate the effectiveness of our method.
- Score: 48.24479287526052
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Accurate customer lifetime value (LTV) prediction can help service providers
optimize their marketing policies in customer-centric applications. However,
the heavy sparsity of consumption events and the interference of data variance
and noise obstruct LTV estimation. Many existing LTV prediction methods
directly train a single-view LTV predictor on consumption samples, which may
yield inaccurate and even biased knowledge extraction. In this paper, we
propose a contrastive multi-view framework for LTV prediction, which is a
plug-and-play solution compatible with various backbone models. It synthesizes
multiple heterogeneous LTV regressors with complementary knowledge to improve
model robustness and captures sample relatedness via contrastive learning to
mitigate the dependency on data abundance. Concretely, we use a decomposed
scheme that converts the LTV prediction problem into a combination of
estimating consumption probability and payment amount. To alleviate the impact
of noisy data on model learning, we propose a multi-view framework that jointly
optimizes multiple types of regressors with diverse characteristics and
advantages to encode and fuse comprehensive knowledge. To fully exploit the
potential of limited training samples, we propose a hybrid contrastive learning
method to help capture the relatedness between samples in both classification
and regression tasks. We conduct extensive experiments on a real-world game LTV
prediction dataset and the results validate the effectiveness of our method. We
have deployed our solution online in Huawei's mobile game center and achieved
32.26% of total payment amount gains.
Related papers
- Scaling Laws for Predicting Downstream Performance in LLMs [75.28559015477137]
This work focuses on the pre-training loss as a more-efficient metric for performance estimation.
We extend the power law analytical function to predict domain-specific pre-training loss based on FLOPs across data sources.
We employ a two-layer neural network to model the non-linear relationship between multiple domain-specific loss and downstream performance.
arXiv Detail & Related papers (2024-10-11T04:57:48Z) - TiMix: Text-aware Image Mixing for Effective Vision-Language
Pre-training [42.142924806184425]
Mixed data samples for cross-modal contrastive learning implicitly serve as a regularizer for the contrastive loss.
TiMix exhibits a comparable performance on downstream tasks, even with a reduced amount of training data and shorter training time, when benchmarked against existing methods.
arXiv Detail & Related papers (2023-12-14T12:02:24Z) - Debiasing Multimodal Models via Causal Information Minimization [65.23982806840182]
We study bias arising from confounders in a causal graph for multimodal data.
Robust predictive features contain diverse information that helps a model generalize to out-of-distribution data.
We use these features as confounder representations and use them via methods motivated by causal theory to remove bias from models.
arXiv Detail & Related papers (2023-11-28T16:46:14Z) - Billion-user Customer Lifetime Value Prediction: An Industrial-scale
Solution from Kuaishou [19.31651596803956]
Customer Life Time Value (LTV) is the expected total revenue that a single user can bring to a business.
Modeling LTV is a challenging problem, due to its complex and mutable data distribution.
We introduce an Order Dependency Monotonic Network (ODMN) that models the ordered dependencies between LTVs of different time spans.
arXiv Detail & Related papers (2022-08-29T04:05:21Z) - DRFLM: Distributionally Robust Federated Learning with Inter-client
Noise via Local Mixup [58.894901088797376]
federated learning has emerged as a promising approach for training a global model using data from multiple organizations without leaking their raw data.
We propose a general framework to solve the above two challenges simultaneously.
We provide comprehensive theoretical analysis including robustness analysis, convergence analysis, and generalization ability.
arXiv Detail & Related papers (2022-04-16T08:08:29Z) - Multitask Learning and Bandits via Robust Statistics [3.103098467546532]
Decision-makers often simultaneously face many related but heterogeneous learning problems.
We propose a novel two-stage multitask learning estimator that exploits this structure in a sample-efficient way.
Our estimator yields improved sample complexity bounds in the feature dimension $d$ relative to commonly-employed estimators.
arXiv Detail & Related papers (2021-12-28T17:37:08Z) - Cooperative learning for multi-view analysis [2.368995563245609]
We propose a new method for supervised learning with multiple sets of features ("views")
Cooperative learning combines the usual squared error loss of predictions with an "agreement" penalty to encourage the predictions from different data views to agree.
We illustrate the effectiveness of our proposed method on simulated and real data examples.
arXiv Detail & Related papers (2021-12-23T03:13:25Z) - Rethinking InfoNCE: How Many Negative Samples Do You Need? [54.146208195806636]
We study how many negative samples are optimal for InfoNCE in different scenarios via a semi-quantitative theoretical framework.
We estimate the optimal negative sampling ratio using the $K$ value that maximizes the training effectiveness function.
arXiv Detail & Related papers (2021-05-27T08:38:29Z) - Learning Diverse Representations for Fast Adaptation to Distribution
Shift [78.83747601814669]
We present a method for learning multiple models, incorporating an objective that pressures each to learn a distinct way to solve the task.
We demonstrate our framework's ability to facilitate rapid adaptation to distribution shift.
arXiv Detail & Related papers (2020-06-12T12:23:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.