MAPLE: Mobile App Prediction Leveraging Large Language Model Embeddings
- URL: http://arxiv.org/abs/2309.08648v3
- Date: Wed, 31 Jan 2024 02:36:48 GMT
- Title: MAPLE: Mobile App Prediction Leveraging Large Language Model Embeddings
- Authors: Yonchanok Khaokaew, Hao Xue, Flora D. Salim
- Abstract summary: This study introduces a novel prediction model, Mobile App Prediction Leveraging Large Language Model Embeddings (MAPLE)
MAPLE employs Large Language Models (LLMs) and installed app similarity to overcome these challenges.
In tests on two real-world datasets, MAPLE surpasses contemporary models in both standard and cold start scenarios.
- Score: 10.15489740838546
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, predicting mobile app usage has become increasingly
important for areas like app recommendation, user behaviour analysis, and
mobile resource management. Existing models, however, struggle with the
heterogeneous nature of contextual data and the user cold start problem. This
study introduces a novel prediction model, Mobile App Prediction Leveraging
Large Language Model Embeddings (MAPLE), which employs Large Language Models
(LLMs) and installed app similarity to overcome these challenges. MAPLE
utilises the power of LLMs to process contextual data and discern intricate
relationships within it effectively. Additionally, we explore the use of
installed app similarity to address the cold start problem, facilitating the
modelling of user preferences and habits, even for new users with limited
historical data. In essence, our research presents MAPLE as a novel, potent,
and practical approach to app usage prediction, making significant strides in
resolving issues faced by existing models. MAPLE stands out as a comprehensive
and effective solution, setting a new benchmark for more precise and
personalised app usage predictions. In tests on two real-world datasets, MAPLE
surpasses contemporary models in both standard and cold start scenarios. These
outcomes validate MAPLE's capacity for precise app usage predictions and its
resilience against the cold start problem. This enhanced performance stems from
the model's proficiency in capturing complex temporal patterns and leveraging
contextual information. As a result, MAPLE can potentially improve personalised
mobile app usage predictions and user experiences markedly.
Related papers
- New User Event Prediction Through the Lens of Causal Inference [20.676353189313737]
We propose a novel discrete event prediction framework for new users.
Our method offers an unbiased prediction for new users without needing to know their categories.
We demonstrate the superior performance of the proposed framework with a numerical simulation study and two real-world applications.
arXiv Detail & Related papers (2024-07-08T05:35:54Z) - Predictive Churn with the Set of Good Models [64.05949860750235]
We study the effect of conflicting predictions over the set of near-optimal machine learning models.
We present theoretical results on the expected churn between models within the Rashomon set.
We show how our approach can be used to better anticipate, reduce, and avoid churn in consumer-facing applications.
arXiv Detail & Related papers (2024-02-12T16:15:25Z) - Federated Privacy-preserving Collaborative Filtering for On-Device Next
App Prediction [52.16923290335873]
We propose a novel SeqMF model to solve the problem of predicting the next app launch during mobile device usage.
We modify the structure of the classical matrix factorization model and update the training procedure to sequential learning.
One more ingredient of the proposed approach is a new privacy mechanism that guarantees the protection of the sent data from the users to the remote server.
arXiv Detail & Related papers (2023-02-05T10:29:57Z) - Synthetic Model Combination: An Instance-wise Approach to Unsupervised
Ensemble Learning [92.89846887298852]
Consider making a prediction over new test data without any opportunity to learn from a training set of labelled data.
Give access to a set of expert models and their predictions alongside some limited information about the dataset used to train them.
arXiv Detail & Related papers (2022-10-11T10:20:31Z) - MRCLens: an MRC Dataset Bias Detection Toolkit [82.44296974850639]
We introduce MRCLens, a toolkit that detects whether biases exist before users train the full model.
For the convenience of introducing the toolkit, we also provide a categorization of common biases in MRC.
arXiv Detail & Related papers (2022-07-18T21:05:39Z) - A Graph-Enhanced Click Model for Web Search [67.27218481132185]
We propose a novel graph-enhanced click model (GraphCM) for web search.
We exploit both intra-session and inter-session information for the sparsity and cold-start problems.
arXiv Detail & Related papers (2022-06-17T08:32:43Z) - Super-Prompting: Utilizing Model-Independent Contextual Data to Reduce
Data Annotation Required in Visual Commonsense Tasks [3.42658286826597]
We analyze different prompt-based fine-tuning techniques to improve results on both language and multimodal causal transformer models.
Our results show that by simple model-agnostic prompt-based fine-tuning, comparable results can be reached by only using 35%-40% of the fine-tuning training dataset.
arXiv Detail & Related papers (2022-04-25T18:56:55Z) - CoSEM: Contextual and Semantic Embedding for App Usage Prediction [8.78798600901882]
Existing modeling approaches utilize historical app usage logs along with a wide range of semantic information to predict the app usage.
This paper develop a model called Contextual and Semantic Embedding model for app usage prediction.
Experiments show that the combination of semantic information and history app usage information enables our model to outperform the baselines.
arXiv Detail & Related papers (2021-08-26T02:38:44Z) - Back2Future: Leveraging Backfill Dynamics for Improving Real-time
Predictions in Future [73.03458424369657]
In real-time forecasting in public health, data collection is a non-trivial and demanding task.
'Backfill' phenomenon and its effect on model performance has been barely studied in the prior literature.
We formulate a novel problem and neural framework Back2Future that aims to refine a given model's predictions in real-time.
arXiv Detail & Related papers (2021-06-08T14:48:20Z) - Privacy Enhancing Machine Learning via Removal of Unwanted Dependencies [21.97951347784442]
This paper studies new variants of supervised and adversarial learning methods, which remove the sensitive information in the data before they are sent out for a particular application.
The explored methods optimize privacy preserving feature mappings and predictive models simultaneously in an end-to-end fashion.
Experimental results on mobile sensing and face datasets demonstrate that our models can successfully maintain the utility performances of predictive models while causing sensitive predictions to perform poorly.
arXiv Detail & Related papers (2020-07-30T19:55:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.