On the Theories Behind Hard Negative Sampling for Recommendation
- URL: http://arxiv.org/abs/2302.03472v1
- Date: Tue, 7 Feb 2023 13:57:03 GMT
- Title: On the Theories Behind Hard Negative Sampling for Recommendation
- Authors: Wentao Shi, Jiawei Chen, Fuli Feng, Jizhi Zhang, Junkang Wu, Chongming
Gao and Xiangnan He
- Abstract summary: We offer two insightful guidelines for effective usage of Hard Negative Sampling (HNS)
We prove that employing HNS on the Personalized Ranking (BPR) learner is equivalent to optimizing One-way Partial AUC (OPAUC)
These analyses establish the theoretical foundation of HNS in optimizing Top-K recommendation performance for the first time.
- Score: 51.64626293229085
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Negative sampling has been heavily used to train recommender models on
large-scale data, wherein sampling hard examples usually not only accelerates
the convergence but also improves the model accuracy. Nevertheless, the reasons
for the effectiveness of Hard Negative Sampling (HNS) have not been revealed
yet. In this work, we fill the research gap by conducting thorough theoretical
analyses on HNS. Firstly, we prove that employing HNS on the Bayesian
Personalized Ranking (BPR) learner is equivalent to optimizing One-way Partial
AUC (OPAUC). Concretely, the BPR equipped with Dynamic Negative Sampling (DNS)
is an exact estimator, while with softmax-based sampling is a soft estimator.
Secondly, we prove that OPAUC has a stronger connection with Top-K evaluation
metrics than AUC and verify it with simulation experiments. These analyses
establish the theoretical foundation of HNS in optimizing Top-K recommendation
performance for the first time. On these bases, we offer two insightful
guidelines for effective usage of HNS: 1) the sampling hardness should be
controllable, e.g., via pre-defined hyper-parameters, to adapt to different
Top-K metrics and datasets; 2) the smaller the $K$ we emphasize in Top-K
evaluation metrics, the harder the negative samples we should draw. Extensive
experiments on three real-world benchmarks verify the two guidelines.
Related papers
- Neural Importance Sampling for Rapid and Reliable Gravitational-Wave
Inference [59.040209568168436]
We first generate a rapid proposal for the Bayesian posterior using neural networks, and then attach importance weights based on the underlying likelihood and prior.
This provides (1) a corrected posterior free from network inaccuracies, (2) a performance diagnostic (the sample efficiency) for assessing the proposal and identifying failure cases, and (3) an unbiased estimate of the Bayesian evidence.
We carry out a large study analyzing 42 binary black hole mergers observed by LIGO and Virgo with the SEOBNRv4PHM and IMRPhenomHMXP waveform models.
arXiv Detail & Related papers (2022-10-11T18:00:02Z) - Ordinal Regression via Binary Preference vs Simple Regression:
Statistical and Experimental Perspectives [11.969202112011963]
Ordinal regression with anchored reference samples (ORARS) has been proposed for predicting the subjective Mean Opinion Score (MOS) of input stimuli automatically.
A trained binary classifier is then used to predict which sample, test or anchor, is better statistically.
arXiv Detail & Related papers (2022-07-06T05:52:19Z) - Insights into performance evaluation of com-pound-protein interaction
prediction methods [0.0]
Machine learning based prediction of compound-protein interactions (CPIs) is important for drug design, screening and repurposing studies.
We have observed a number of fundamental issues in experiment design that lead to over optimistic estimates of model performance.
arXiv Detail & Related papers (2022-01-28T20:07:19Z) - Sensing Cox Processes via Posterior Sampling and Positive Bases [56.82162768921196]
We study adaptive sensing of point processes, a widely used model from spatial statistics.
We model the intensity function as a sample from a truncated Gaussian process, represented in a specially constructed positive basis.
Our adaptive sensing algorithms use Langevin dynamics and are based on posterior sampling (textscCox-Thompson) and top-two posterior sampling (textscTop2) principles.
arXiv Detail & Related papers (2021-10-21T14:47:06Z) - Towards the Memorization Effect of Neural Networks in Adversarial
Training [36.35802928128805]
We propose Benign Adversarial Training (BAT) which can facilitate adversarial training to avoid fitting harmful'' atypical samples.
BAT can achieve better clean accuracy vs. robustness trade-off than baseline methods, in benchmark datasets such as CIFAR100 and TinyImageNet.
arXiv Detail & Related papers (2021-06-09T03:47:32Z) - Rethinking InfoNCE: How Many Negative Samples Do You Need? [54.146208195806636]
We study how many negative samples are optimal for InfoNCE in different scenarios via a semi-quantitative theoretical framework.
We estimate the optimal negative sampling ratio using the $K$ value that maximizes the training effectiveness function.
arXiv Detail & Related papers (2021-05-27T08:38:29Z) - Understanding Negative Sampling in Graph Representation Learning [87.35038268508414]
We show that negative sampling is as important as positive sampling in determining the optimization objective and the resulted variance.
We propose Metropolis-Hastings (MCNS) to approximate the positive distribution with self-contrast approximation and accelerate negative sampling by Metropolis-Hastings.
We evaluate our method on 5 datasets that cover extensive downstream graph learning tasks, including link prediction, node classification and personalized recommendation.
arXiv Detail & Related papers (2020-05-20T06:25:21Z) - Reinforced Negative Sampling over Knowledge Graph for Recommendation [106.07209348727564]
We develop a new negative sampling model, Knowledge Graph Policy Network (kgPolicy), which works as a reinforcement learning agent to explore high-quality negatives.
kgPolicy navigates from the target positive interaction, adaptively receives knowledge-aware negative signals, and ultimately yields a potential negative item to train the recommender.
arXiv Detail & Related papers (2020-03-12T12:44:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.