SimpleX: A Simple and Strong Baseline for Collaborative Filtering
- URL: http://arxiv.org/abs/2109.12613v3
- Date: Thu, 30 Nov 2023 02:38:32 GMT
- Title: SimpleX: A Simple and Strong Baseline for Collaborative Filtering
- Authors: Kelong Mao, Jieming Zhu, Jinpeng Wang, Quanyu Dai, Zhenhua Dong, Xi
Xiao, Xiuqiang He
- Abstract summary: Collaborative filtering (CF) is a widely studied research topic in recommender systems.
We show that the choice of loss function as well as negative sampling ratio is equivalently important.
We propose the cosine contrastive loss (CCL) and further incorporate it to a simple unified CF model, dubbed SimpleX.
- Score: 50.30070461560722
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Collaborative filtering (CF) is a widely studied research topic in
recommender systems. The learning of a CF model generally depends on three
major components, namely interaction encoder, loss function, and negative
sampling. While many existing studies focus on the design of more powerful
interaction encoders, the impacts of loss functions and negative sampling
ratios have not yet been well explored. In this work, we show that the choice
of loss function as well as negative sampling ratio is equivalently important.
More specifically, we propose the cosine contrastive loss (CCL) and further
incorporate it to a simple unified CF model, dubbed SimpleX. Extensive
experiments have been conducted on 11 benchmark datasets and compared with 29
existing CF models in total. Surprisingly, the results show that, under our CCL
loss and a large negative sampling ratio, SimpleX can surpass most
sophisticated state-of-the-art models by a large margin (e.g., max 48.5%
improvement in NDCG@20 over LightGCN). We believe that SimpleX could not only
serve as a simple strong baseline to foster future research on CF, but also
shed light on the potential research direction towards improving loss function
and negative sampling. Our source code will be available at
https://reczoo.github.io/SimpleX.
Related papers
- Adversarial Collaborative Filtering for Free [27.949683060138064]
Collaborative Filtering (CF) has been successfully used to help users discover the items of interest.
Existing methods suffer from noisy data issue, which negatively impacts the quality of recommendation.
We present Sharpness-aware Collaborative Filtering (CF), a simple yet effective method that conducts adversarial training without extra computational cost over the base.
arXiv Detail & Related papers (2023-08-20T19:25:38Z) - uCTRL: Unbiased Contrastive Representation Learning via Alignment and
Uniformity for Collaborative Filtering [6.663503238373593]
Collaborative filtering (CF) models tend to yield recommendation lists with popularity bias.
We propose Unbiased ConTrastive Representation Learning (uCTRL) to mitigate this problem.
We also devise a novel IPW estimation method that removes the bias of both users and items.
arXiv Detail & Related papers (2023-05-22T06:55:38Z) - Siamese Prototypical Contrastive Learning [24.794022951873156]
Contrastive Self-supervised Learning (CSL) is a practical solution that learns meaningful visual representations from massive data in an unsupervised approach.
In this paper, we tackle this problem by introducing a simple but effective contrastive learning framework.
The key insight is to employ siamese-style metric loss to match intra-prototype features, while increasing the distance between inter-prototype features.
arXiv Detail & Related papers (2022-08-18T13:25:30Z) - SelfCF: A Simple Framework for Self-supervised Collaborative Filtering [72.68215241599509]
Collaborative filtering (CF) is widely used to learn informative latent representations of users and items from observed interactions.
We propose a self-supervised collaborative filtering framework (SelfCF) that is specially designed for recommender scenario with implicit feedback.
We show that SelfCF can boost up the accuracy by up to 17.79% on average, compared with a self-supervised framework BUIR.
arXiv Detail & Related papers (2021-07-07T05:21:12Z) - Examining and Combating Spurious Features under Distribution Shift [94.31956965507085]
We define and analyze robust and spurious representations using the information-theoretic concept of minimal sufficient statistics.
We prove that even when there is only bias of the input distribution, models can still pick up spurious features from their training data.
Inspired by our analysis, we demonstrate that group DRO can fail when groups do not directly account for various spurious correlations.
arXiv Detail & Related papers (2021-06-14T05:39:09Z) - Contrastive Attraction and Contrastive Repulsion for Representation
Learning [131.72147978462348]
Contrastive learning (CL) methods learn data representations in a self-supervision manner, where the encoder contrasts each positive sample over multiple negative samples.
Recent CL methods have achieved promising results when pretrained on large-scale datasets, such as ImageNet.
We propose a doubly CL strategy that separately compares positive and negative samples within their own groups, and then proceeds with a contrast between positive and negative groups.
arXiv Detail & Related papers (2021-05-08T17:25:08Z) - Enhanced Principal Component Analysis under A Collaborative-Robust
Framework [89.28334359066258]
We introduce a general collaborative-robust weight learning framework that combines weight learning and robust loss in a non-trivial way.
Under the proposed framework, only a part of well-fitting samples are activated which indicates more importance during training, and others, whose errors are large, will not be ignored.
In particular, the negative effects of inactivated samples are alleviated by the robust loss function.
arXiv Detail & Related papers (2021-03-22T15:17:37Z) - EqCo: Equivalent Rules for Self-supervised Contrastive Learning [81.45848885547754]
We propose a method to make self-supervised learning irrelevant to the number of negative samples in InfoNCE-based contrastive learning frameworks.
Inspired by the InfoMax principle, we point that the margin term in contrastive loss needs to be adaptively scaled according to the number of negative pairs.
arXiv Detail & Related papers (2020-10-05T11:39:04Z) - SCE: Scalable Network Embedding from Sparsest Cut [20.08464038805681]
Large-scale network embedding is to learn a latent representation for each node in an unsupervised manner.
A key of success to such contrastive learning methods is how to draw positive and negative samples.
In this paper, we propose SCE for unsupervised network embedding only using negative samples for training.
arXiv Detail & Related papers (2020-06-30T03:18:15Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.