Dimension Independent Mixup for Hard Negative Sample in Collaborative
Filtering
- URL: http://arxiv.org/abs/2306.15905v2
- Date: Fri, 18 Aug 2023 04:24:24 GMT
- Title: Dimension Independent Mixup for Hard Negative Sample in Collaborative
Filtering
- Authors: Xi Wu, Liangwei Yang, Jibing Gong, Chao Zhou, Tianyu Lin, Xiaolong
Liu, Philip S. Yu
- Abstract summary: Negative sampling plays a vital role in training CF-based models with implicit feedback.
We propose Dimension Independent Mixup for Hard Negative Sampling (DINS), which is the first Area-wise sampling method for training CF-based models.
Our work contributes a new perspective, introduces Area-wise sampling, and presents DINS as a novel approach for negative sampling.
- Score: 36.26865960551565
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Collaborative filtering (CF) is a widely employed technique that predicts
user preferences based on past interactions. Negative sampling plays a vital
role in training CF-based models with implicit feedback. In this paper, we
propose a novel perspective based on the sampling area to revisit existing
sampling methods. We point out that current sampling methods mainly focus on
Point-wise or Line-wise sampling, lacking flexibility and leaving a significant
portion of the hard sampling area un-explored. To address this limitation, we
propose Dimension Independent Mixup for Hard Negative Sampling (DINS), which is
the first Area-wise sampling method for training CF-based models. DINS
comprises three modules: Hard Boundary Definition, Dimension Independent Mixup,
and Multi-hop Pooling. Experiments with real-world datasets on both matrix
factorization and graph-based models demonstrate that DINS outperforms other
negative sampling methods, establishing its effectiveness and superiority. Our
work contributes a new perspective, introduces Area-wise sampling, and presents
DINS as a novel approach that achieves state-of-the-art performance for
negative sampling. Our implementations are available in PyTorch.
Related papers
- Your Diffusion Model is Secretly a Noise Classifier and Benefits from Contrastive Training [20.492630610281658]
Diffusion models learn to denoise data and the trained denoiser is then used to generate new samples from the data distribution.
We introduce a new self-supervised training objective that differentiates the levels of noise added to a sample.
We show by diverse experiments that the proposed contrastive diffusion training is effective for both sequential and parallel settings.
arXiv Detail & Related papers (2024-07-12T03:03:50Z) - A Diffusion Model Framework for Unsupervised Neural Combinatorial Optimization [7.378582040635655]
Current deep learning approaches rely on generative models that yield exact sample likelihoods.
This work introduces a method that lifts this restriction and opens the possibility to employ highly expressive latent variable models.
We experimentally validate our approach in data-free Combinatorial Optimization and demonstrate that our method achieves a new state-of-the-art on a wide range of benchmark problems.
arXiv Detail & Related papers (2024-06-03T17:55:02Z) - Stochastic Sampling for Contrastive Views and Hard Negative Samples in Graph-based Collaborative Filtering [28.886714896469737]
Graph-based collaborative filtering (CF) has emerged as a promising approach in recommendation systems.
Despite its achievements, graph-based CF models face challenges due to data sparsity and negative sampling.
We propose a novel sampling for i) COntrastive views and ii) hard NEgative samples (SCONE) to overcome these issues.
arXiv Detail & Related papers (2024-05-01T02:27:59Z) - Improved off-policy training of diffusion samplers [93.66433483772055]
We study the problem of training diffusion models to sample from a distribution with an unnormalized density or energy function.
We benchmark several diffusion-structured inference methods, including simulation-based variational approaches and off-policy methods.
Our results shed light on the relative advantages of existing algorithms while bringing into question some claims from past work.
arXiv Detail & Related papers (2024-02-07T18:51:49Z) - Improved Active Learning via Dependent Leverage Score Sampling [8.400581768343804]
We show how to obtain improved active learning methods in the agnostic (adversarial noise) setting.
We propose an easily implemented method based on the emphpivotal sampling algorithm
In comparison to independent sampling, our method reduces the number of samples needed to reach a given target accuracy by up to $50%$.
arXiv Detail & Related papers (2023-10-08T01:51:30Z) - Rethinking Collaborative Metric Learning: Toward an Efficient
Alternative without Negative Sampling [156.7248383178991]
Collaborative Metric Learning (CML) paradigm has aroused wide interest in the area of recommendation systems (RS)
We find that negative sampling would lead to a biased estimation of the generalization error.
Motivated by this, we propose an efficient alternative without negative sampling for CML named textitSampling-Free Collaborative Metric Learning (SFCML)
arXiv Detail & Related papers (2022-06-23T08:50:22Z) - Unrolling Particles: Unsupervised Learning of Sampling Distributions [102.72972137287728]
Particle filtering is used to compute good nonlinear estimates of complex systems.
We show in simulations that the resulting particle filter yields good estimates in a wide range of scenarios.
arXiv Detail & Related papers (2021-10-06T16:58:34Z) - Beyond Farthest Point Sampling in Point-Wise Analysis [52.218037492342546]
We propose a novel data-driven sampler learning strategy for point-wise analysis tasks.
We learn sampling and downstream applications jointly.
Our experiments show that jointly learning of the sampler and task brings remarkable improvement over previous baseline methods.
arXiv Detail & Related papers (2021-07-09T08:08:44Z) - Jo-SRC: A Contrastive Approach for Combating Noisy Labels [58.867237220886885]
We propose a noise-robust approach named Jo-SRC (Joint Sample Selection and Model Regularization based on Consistency)
Specifically, we train the network in a contrastive learning manner. Predictions from two different views of each sample are used to estimate its "likelihood" of being clean or out-of-distribution.
arXiv Detail & Related papers (2021-03-24T07:26:07Z) - Understanding Negative Sampling in Graph Representation Learning [87.35038268508414]
We show that negative sampling is as important as positive sampling in determining the optimization objective and the resulted variance.
We propose Metropolis-Hastings (MCNS) to approximate the positive distribution with self-contrast approximation and accelerate negative sampling by Metropolis-Hastings.
We evaluate our method on 5 datasets that cover extensive downstream graph learning tasks, including link prediction, node classification and personalized recommendation.
arXiv Detail & Related papers (2020-05-20T06:25:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.