Learning Explicit User Interest Boundary for Recommendation
- URL: http://arxiv.org/abs/2111.11026v1
- Date: Mon, 22 Nov 2021 07:26:51 GMT
- Title: Learning Explicit User Interest Boundary for Recommendation
- Authors: Jianhuan Zhuo, Qiannan Zhu, Yinliang Yue and Yuhong Zhao
- Abstract summary: We introduce an auxiliary score $b_u$ for each user to represent the User Interest Boundary.
We show that our approach can provide a personalized decision boundary and significantly improve the training efficiency without any special sampling strategy.
- Score: 5.715918678913698
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: The core objective of modelling recommender systems from implicit feedback is
to maximize the positive sample score $s_p$ and minimize the negative sample
score $s_n$, which can usually be summarized into two paradigms: the pointwise
and the pairwise. The pointwise approaches fit each sample with its label
individually, which is flexible in weighting and sampling on instance-level but
ignores the inherent ranking property. By qualitatively minimizing the relative
score $s_n - s_p$, the pairwise approaches capture the ranking of samples
naturally but suffer from training efficiency. Additionally, both approaches
are hard to explicitly provide a personalized decision boundary to determine if
users are interested in items unseen. To address those issues, we innovatively
introduce an auxiliary score $b_u$ for each user to represent the User Interest
Boundary(UIB) and individually penalize samples that cross the boundary with
pairwise paradigms, i.e., the positive samples whose score is lower than $b_u$
and the negative samples whose score is higher than $b_u$. In this way, our
approach successfully achieves a hybrid loss of the pointwise and the pairwise
to combine the advantages of both. Analytically, we show that our approach can
provide a personalized decision boundary and significantly improve the training
efficiency without any special sampling strategy. Extensive results show that
our approach achieves significant improvements on not only the classical
pointwise or pairwise models but also state-of-the-art models with complex loss
function and complicated feature encoding.
Related papers
- Task-oriented Embedding Counts: Heuristic Clustering-driven Feature Fine-tuning for Whole Slide Image Classification [1.292108130501585]
We propose a clustering-driven feature fine-tuning method (HC-FT) to enhance the performance of multiple instance learning.
The proposed method is evaluated on both CAMELYON16 and BRACS datasets, achieving an AUC of 97.13% and 85.85%, respectively.
arXiv Detail & Related papers (2024-06-02T08:53:45Z) - CKD: Contrastive Knowledge Distillation from A Sample-wise Perspective [48.99488315273868]
We present a contrastive knowledge distillation approach, which can be formulated as a sample-wise alignment problem with intra- and inter-sample constraints.
Our method minimizes logit differences within the same sample by considering their numerical values.
We conduct comprehensive experiments on three datasets including CIFAR-100, ImageNet-1K, and MS COCO.
arXiv Detail & Related papers (2024-04-22T11:52:40Z) - Improved Active Learning via Dependent Leverage Score Sampling [8.400581768343804]
We show how to obtain improved active learning methods in the agnostic (adversarial noise) setting.
We propose an easily implemented method based on the emphpivotal sampling algorithm
In comparison to independent sampling, our method reduces the number of samples needed to reach a given target accuracy by up to $50%$.
arXiv Detail & Related papers (2023-10-08T01:51:30Z) - Rethinking Clustering-Based Pseudo-Labeling for Unsupervised
Meta-Learning [146.11600461034746]
Method for unsupervised meta-learning, CACTUs, is a clustering-based approach with pseudo-labeling.
This approach is model-agnostic and can be combined with supervised algorithms to learn from unlabeled data.
We prove that the core reason for this is lack of a clustering-friendly property in the embedding space.
arXiv Detail & Related papers (2022-09-27T19:04:36Z) - Rethinking InfoNCE: How Many Negative Samples Do You Need? [54.146208195806636]
We study how many negative samples are optimal for InfoNCE in different scenarios via a semi-quantitative theoretical framework.
We estimate the optimal negative sampling ratio using the $K$ value that maximizes the training effectiveness function.
arXiv Detail & Related papers (2021-05-27T08:38:29Z) - Scalable Personalised Item Ranking through Parametric Density Estimation [53.44830012414444]
Learning from implicit feedback is challenging because of the difficult nature of the one-class problem.
Most conventional methods use a pairwise ranking approach and negative samplers to cope with the one-class problem.
We propose a learning-to-rank approach, which achieves convergence speed comparable to the pointwise counterpart.
arXiv Detail & Related papers (2021-05-11T03:38:16Z) - Doubly Contrastive Deep Clustering [135.7001508427597]
We present a novel Doubly Contrastive Deep Clustering (DCDC) framework, which constructs contrastive loss over both sample and class views.
Specifically, for the sample view, we set the class distribution of the original sample and its augmented version as positive sample pairs.
For the class view, we build the positive and negative pairs from the sample distribution of the class.
In this way, two contrastive losses successfully constrain the clustering results of mini-batch samples in both sample and class level.
arXiv Detail & Related papers (2021-03-09T15:15:32Z) - Multi-label Contrastive Predictive Coding [125.03510235962095]
Variational mutual information (MI) estimators are widely used in unsupervised representation learning methods such as contrastive predictive coding (CPC)
We introduce a novel estimator based on a multi-label classification problem, where the critic needs to jointly identify multiple positive samples at the same time.
We show that using the same amount of negative samples, multi-label CPC is able to exceed the $log m$ bound, while still being a valid lower bound of mutual information.
arXiv Detail & Related papers (2020-07-20T02:46:21Z) - Addressing Class-Imbalance Problem in Personalized Ranking [47.11372043636176]
We propose an efficient emphunderlineVital underlineNegative underlineSampler (VINS) to alleviate the class-imbalance issue for pairwise ranking model.
VINS is a bias sampler with reject probability that will tend to accept a negative candidate with a larger degree weight than the given positive item.
arXiv Detail & Related papers (2020-05-19T08:11:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.