Billion-user Customer Lifetime Value Prediction: An Industrial-scale
Solution from Kuaishou
- URL: http://arxiv.org/abs/2208.13358v1
- Date: Mon, 29 Aug 2022 04:05:21 GMT
- Title: Billion-user Customer Lifetime Value Prediction: An Industrial-scale
Solution from Kuaishou
- Authors: Kunpeng Li, Guangcui Shao, Naijun Yang, Xiao Fang, Yang Song
- Abstract summary: Customer Life Time Value (LTV) is the expected total revenue that a single user can bring to a business.
Modeling LTV is a challenging problem, due to its complex and mutable data distribution.
We introduce an Order Dependency Monotonic Network (ODMN) that models the ordered dependencies between LTVs of different time spans.
- Score: 19.31651596803956
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Customer Life Time Value (LTV) is the expected total revenue that a single
user can bring to a business. It is widely used in a variety of business
scenarios to make operational decisions when acquiring new customers. Modeling
LTV is a challenging problem, due to its complex and mutable data distribution.
Existing approaches either directly learn from posterior feature distributions
or leverage statistical models that make strong assumption on prior
distributions, both of which fail to capture those mutable distributions. In
this paper, we propose a complete set of industrial-level LTV modeling
solutions. Specifically, we introduce an Order Dependency Monotonic Network
(ODMN) that models the ordered dependencies between LTVs of different time
spans, which greatly improves model performance. We further introduce a Multi
Distribution Multi Experts (MDME) module based on the Divide-and-Conquer idea,
which transforms the severely imbalanced distribution modeling problem into a
series of relatively balanced sub-distribution modeling problems hence greatly
reduces the modeling complexity. In addition, a novel evaluation metric Mutual
Gini is introduced to better measure the distribution difference between the
estimated value and the ground-truth label based on the Lorenz Curve. The ODMN
framework has been successfully deployed in many business scenarios of
Kuaishou, and achieved great performance. Extensive experiments on real-world
industrial data demonstrate the superiority of the proposed methods compared to
state-of-the-art baselines including ZILN and Two-Stage XGBoost models.
Related papers
- MoD: A Distribution-Based Approach for Merging Large Language Models [0.0]
Large language models (LLMs) have enabled the development of numerous specialized, task-specific variants.
We propose the textitMixture of Distributions (MoD) framework, a novel approach for merging LLMs.
Unlike traditional weight-averaging methods, MoD effectively preserves the specialized capabilities of individual models.
arXiv Detail & Related papers (2024-11-01T07:05:29Z) - xGen-MM (BLIP-3): A Family of Open Large Multimodal Models [157.44696790158784]
This report introduces xGen-MM, a framework for developing Large Multimodal Models (LMMs)
The framework comprises meticulously curated datasets, a training recipe, model architectures, and a resulting suite of LMMs.
Our models undergo rigorous evaluation across a range of tasks, including both single and multi-image benchmarks.
arXiv Detail & Related papers (2024-08-16T17:57:01Z) - MAP: Low-compute Model Merging with Amortized Pareto Fronts via Quadratic Approximation [80.47072100963017]
We introduce a novel and low-compute algorithm, Model Merging with Amortized Pareto Front (MAP)
MAP efficiently identifies a set of scaling coefficients for merging multiple models, reflecting the trade-offs involved.
We also introduce Bayesian MAP for scenarios with a relatively low number of tasks and Nested MAP for situations with a high number of tasks, further reducing the computational cost of evaluation.
arXiv Detail & Related papers (2024-06-11T17:55:25Z) - EMR-Merging: Tuning-Free High-Performance Model Merging [55.03509900949149]
We show that Elect, Mask & Rescale-Merging (EMR-Merging) shows outstanding performance compared to existing merging methods.
EMR-Merging is tuning-free, thus requiring no data availability or any additional training while showing impressive performance.
arXiv Detail & Related papers (2024-05-23T05:25:45Z) - Robust Fine-Tuning of Vision-Language Models for Domain Generalization [6.7181844004432385]
Foundation models have impressive zero-shot inference capabilities and robustness under distribution shifts.
We present a new recipe for few-shot fine-tuning of the popular vision-language foundation model CLIP.
Our experimentation demonstrates that, while zero-shot CLIP fails to match performance of trained vision models on more complex benchmarks, few-shot CLIP fine-tuning outperforms its vision-only counterparts.
arXiv Detail & Related papers (2023-11-03T20:50:40Z) - Adapting Large Language Models for Content Moderation: Pitfalls in Data
Engineering and Supervised Fine-tuning [79.53130089003986]
Large Language Models (LLMs) have become a feasible solution for handling tasks in various domains.
In this paper, we introduce how to fine-tune a LLM model that can be privately deployed for content moderation.
arXiv Detail & Related papers (2023-10-05T09:09:44Z) - Optimal Transport Model Distributional Robustness [33.24747882707421]
Previous works have mainly focused on exploiting distributional robustness in the data space.
We develop theories that enable us to learn the optimal robust center model distribution.
Our framework can be seen as a probabilistic extension of Sharpness-Aware Minimization.
arXiv Detail & Related papers (2023-06-07T06:15:12Z) - Non-Invasive Fairness in Learning through the Lens of Data Drift [88.37640805363317]
We show how to improve the fairness of Machine Learning models without altering the data or the learning algorithm.
We use a simple but key insight: the divergence of trends between different populations, and, consecutively, between a learned model and minority populations, is analogous to data drift.
We explore two strategies (model-splitting and reweighing) to resolve this drift, aiming to improve the overall conformance of models to the underlying data.
arXiv Detail & Related papers (2023-03-30T17:30:42Z) - Deep incremental learning models for financial temporal tabular datasets
with distribution shifts [0.9790236766474201]
The framework uses a simple basic building block (decision trees) to build self-similar models of any required complexity.
We demonstrate our scheme using XGBoost models trained on the Numerai dataset and show that a two layer deep ensemble of XGBoost models over different model snapshots delivers high quality predictions.
arXiv Detail & Related papers (2023-03-14T14:10:37Z) - AvgOut: A Simple Output-Probability Measure to Eliminate Dull Responses [97.50616524350123]
We build dialogue models that are dynamically aware of what utterances or tokens are dull without any feature-engineering.
The first model, MinAvgOut, directly maximizes the diversity score through the output distributions of each batch.
The second model, Label Fine-Tuning (LFT), prepends to the source sequence a label continuously scaled by the diversity score to control the diversity level.
The third model, RL, adopts Reinforcement Learning and treats the diversity score as a reward signal.
arXiv Detail & Related papers (2020-01-15T18:32:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.