Personalizing Intervened Network for Long-tailed Sequential User
Behavior Modeling
- URL: http://arxiv.org/abs/2208.09130v1
- Date: Fri, 19 Aug 2022 02:50:19 GMT
- Title: Personalizing Intervened Network for Long-tailed Sequential User
Behavior Modeling
- Authors: Zheqi Lv, Feng Wang, Shengyu Zhang, Kun Kuang, Hongxia Yang, Fei Wu
- Abstract summary: Tail users suffer from significantly lower-quality recommendation than the head users after joint training.
A model trained on tail users separately still achieve inferior results due to limited data.
We propose a novel approach that significantly improves the recommendation performance of the tail users.
- Score: 66.02953670238647
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In an era of information explosion, recommendation systems play an important
role in people's daily life by facilitating content exploration. It is known
that user activeness, i.e., number of behaviors, tends to follow a long-tail
distribution, where the majority of users are with low activeness. In practice,
we observe that tail users suffer from significantly lower-quality
recommendation than the head users after joint training. We further identify
that a model trained on tail users separately still achieve inferior results
due to limited data. Though long-tail distributions are ubiquitous in
recommendation systems, improving the recommendation performance on the tail
users still remains challenge in both research and industry. Directly applying
related methods on long-tail distribution might be at risk of hurting the
experience of head users, which is less affordable since a small portion of
head users with high activeness contribute a considerate portion of platform
revenue. In this paper, we propose a novel approach that significantly improves
the recommendation performance of the tail users while achieving at least
comparable performance for the head users over the base model. The essence of
this approach is a novel Gradient Aggregation technique that learns common
knowledge shared by all users into a backbone model, followed by separate
plugin prediction networks for the head users and the tail users
personalization. As for common knowledge learning, we leverage the backward
adjustment from the causality theory for deconfounding the gradient estimation
and thus shielding off the backbone training from the confounder, i.e., user
activeness. We conduct extensive experiments on two public recommendation
benchmark datasets and a large-scale industrial datasets collected from the
Alipay platform. Empirical studies validate the rationality and effectiveness
of our approach.
Related papers
- Separating and Learning Latent Confounders to Enhancing User Preferences Modeling [6.0853798070913845]
We propose a novel framework, Separating and Learning Latent Confounders For Recommendation (SLFR)
SLFR obtains the representation of unmeasured confounders to identify the counterfactual feedback by disentangling user preferences and unmeasured confounders.
Experiments in five real-world datasets validate the advantages of our method.
arXiv Detail & Related papers (2023-11-02T08:42:50Z) - Explainable Active Learning for Preference Elicitation [0.0]
We employ Active Learning (AL) to solve the addressed problem with the objective of maximizing information acquisition with minimal user effort.
AL operates for selecting informative data from a large unlabeled set to inquire an oracle to label them.
It harvests user feedback (given for the system's explanations on the presented items) over informative samples to update an underlying machine learning (ML) model.
arXiv Detail & Related papers (2023-09-01T09:22:33Z) - Latent User Intent Modeling for Sequential Recommenders [92.66888409973495]
Sequential recommender models learn to predict the next items a user is likely to interact with based on his/her interaction history on the platform.
Most sequential recommenders however lack a higher-level understanding of user intents, which often drive user behaviors online.
Intent modeling is thus critical for understanding users and optimizing long-term user experience.
arXiv Detail & Related papers (2022-11-17T19:00:24Z) - Constructing Balance from Imbalance for Long-tailed Image Recognition [50.6210415377178]
The imbalance between majority (head) classes and minority (tail) classes severely skews the data-driven deep neural networks.
Previous methods tackle with data imbalance from the viewpoints of data distribution, feature space, and model design.
We propose a concise paradigm by progressively adjusting label space and dividing the head classes and tail classes.
Our proposed model also provides a feature evaluation method and paves the way for long-tailed feature learning.
arXiv Detail & Related papers (2022-08-04T10:22:24Z) - SURF: Semi-supervised Reward Learning with Data Augmentation for
Feedback-efficient Preference-based Reinforcement Learning [168.89470249446023]
We present SURF, a semi-supervised reward learning framework that utilizes a large amount of unlabeled samples with data augmentation.
In order to leverage unlabeled samples for reward learning, we infer pseudo-labels of the unlabeled samples based on the confidence of the preference predictor.
Our experiments demonstrate that our approach significantly improves the feedback-efficiency of the preference-based method on a variety of locomotion and robotic manipulation tasks.
arXiv Detail & Related papers (2022-03-18T16:50:38Z) - Deep Causal Reasoning for Recommendations [47.83224399498504]
A new trend in recommender system research is to negate the influence of confounders from a causal perspective.
We model the recommendation as a multi-cause multi-outcome (MCMO) inference problem.
We show that MCMO modeling may lead to high variance due to scarce observations associated with the high-dimensional causal space.
arXiv Detail & Related papers (2022-01-06T15:00:01Z) - Learning Transferrable Parameters for Long-tailed Sequential User
Behavior Modeling [70.64257515361972]
We argue that focusing on tail users could bring more benefits and address the long tails issue.
Specifically, we propose a gradient alignment and adopt an adversarial training scheme to facilitate knowledge transfer from the head to the tail.
arXiv Detail & Related papers (2020-10-22T03:12:02Z) - Self-Supervised Contrastive Learning for Efficient User Satisfaction
Prediction in Conversational Agents [35.2098736872247]
We propose a self-supervised contrastive learning approach to learn user-agent interactions.
We show that the pre-trained models using the self-supervised objective are transferable to the user satisfaction prediction.
We also propose a novel few-shot transfer learning approach that ensures better transferability for very small sample sizes.
arXiv Detail & Related papers (2020-10-21T18:10:58Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.