PUNR: Pre-training with User Behavior Modeling for News Recommendation
- URL: http://arxiv.org/abs/2304.12633v2
- Date: Mon, 30 Oct 2023 08:13:42 GMT
- Title: PUNR: Pre-training with User Behavior Modeling for News Recommendation
- Authors: Guangyuan Ma, Hongtao Liu, Xing Wu, Wanhui Qian, Zhepeng Lv, Qing
Yang, Songlin Hu
- Abstract summary: News recommendation aims to predict click behaviors based on user behaviors.
How to effectively model the user representations is the key to recommending preferred news.
We propose an unsupervised pre-training paradigm with two tasks, i.e. user behavior masking and user behavior generation.
- Score: 26.349183393252115
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: News recommendation aims to predict click behaviors based on user behaviors.
How to effectively model the user representations is the key to recommending
preferred news. Existing works are mostly focused on improvements in the
supervised fine-tuning stage. However, there is still a lack of PLM-based
unsupervised pre-training methods optimized for user representations. In this
work, we propose an unsupervised pre-training paradigm with two tasks, i.e.
user behavior masking and user behavior generation, both towards effective user
behavior modeling. Firstly, we introduce the user behavior masking pre-training
task to recover the masked user behaviors based on their contextual behaviors.
In this way, the model could capture a much stronger and more comprehensive
user news reading pattern. Besides, we incorporate a novel auxiliary user
behavior generation pre-training task to enhance the user representation vector
derived from the user encoder. We use the above pre-trained user modeling
encoder to obtain news and user representations in downstream fine-tuning.
Evaluations on the real-world news benchmark show significant performance
improvements over existing baselines.
Related papers
- USE: Dynamic User Modeling with Stateful Sequence Models [26.74966828348815]
User Stateful Embedding (USE) generates user embeddings without the need for exhaustive reprocessing.
We introduce a novel training objective named future W-behavior prediction to transcend the limitations of next-token prediction.
We conduct experiments on 8 downstream tasks using Snapchat users' behavioral logs in both static (i.e., fixed user behavior sequences) and dynamic (i.e. periodically updated user behavior sequences) settings.
arXiv Detail & Related papers (2024-03-20T07:05:19Z) - Latent User Intent Modeling for Sequential Recommenders [92.66888409973495]
Sequential recommender models learn to predict the next items a user is likely to interact with based on his/her interaction history on the platform.
Most sequential recommenders however lack a higher-level understanding of user intents, which often drive user behaviors online.
Intent modeling is thus critical for understanding users and optimizing long-term user experience.
arXiv Detail & Related papers (2022-11-17T19:00:24Z) - Meta-Wrapper: Differentiable Wrapping Operator for User Interest
Selection in CTR Prediction [97.99938802797377]
Click-through rate (CTR) prediction, whose goal is to predict the probability of the user to click on an item, has become increasingly significant in recommender systems.
Recent deep learning models with the ability to automatically extract the user interest from his/her behaviors have achieved great success.
We propose a novel approach under the framework of the wrapper method, which is named Meta-Wrapper.
arXiv Detail & Related papers (2022-06-28T03:28:15Z) - Recommender Transformers with Behavior Pathways [50.842316273120744]
We build the Recommender Transformer (RETR) with a novel Pathway Attention mechanism.
We empirically verify the effectiveness of RETR on seven real-world datasets.
arXiv Detail & Related papers (2022-06-13T08:58:37Z) - PinnerFormer: Sequence Modeling for User Representation at Pinterest [60.335384724891746]
We introduce PinnerFormer, a user representation trained to predict a user's future long-term engagement.
Unlike prior approaches, we adapt our modeling to a batch infrastructure via our new dense all-action loss.
We show that by doing so, we significantly close the gap between batch user embeddings that are generated once a day and realtime user embeddings generated whenever a user takes an action.
arXiv Detail & Related papers (2022-05-09T18:26:51Z) - Modeling Dynamic User Preference via Dictionary Learning for Sequential
Recommendation [133.8758914874593]
Capturing the dynamics in user preference is crucial to better predict user future behaviors because user preferences often drift over time.
Many existing recommendation algorithms -- including both shallow and deep ones -- often model such dynamics independently.
This paper considers the problem of embedding a user's sequential behavior into the latent space of user preferences.
arXiv Detail & Related papers (2022-04-02T03:23:46Z) - Generative Adversarial Reward Learning for Generalized Behavior Tendency
Inference [71.11416263370823]
We propose a generative inverse reinforcement learning for user behavioral preference modelling.
Our model can automatically learn the rewards from user's actions based on discriminative actor-critic network and Wasserstein GAN.
arXiv Detail & Related papers (2021-05-03T13:14:25Z) - Reinforced Imitative Graph Representation Learning for Mobile User
Profiling: An Adversarial Training Perspective [21.829562421373712]
We study the problem of mobile user profiling, which is a critical component for quantifying users' characteristics in the human mobility modeling pipeline.
We propose an imitation-based mobile user profiling framework by exploiting reinforcement learning.
arXiv Detail & Related papers (2021-01-07T17:10:00Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.