Semi-supervised Collaborative Filtering by Text-enhanced Domain
Adaptation
- URL: http://arxiv.org/abs/2007.07085v1
- Date: Sun, 28 Jun 2020 05:28:05 GMT
- Title: Semi-supervised Collaborative Filtering by Text-enhanced Domain
Adaptation
- Authors: Wenhui Yu and Xiao Lin and Junfeng Ge and Wenwu Ou and Zheng Qin
- Abstract summary: We consider the problem of recommendation on sparse implicit feedbacks as a semi-supervised learning task.
We focus on the most challenging case -- there is no user or item overlap.
We adopt domain-invariant textual features as the anchor points to align the latent spaces.
- Score: 32.93934837792708
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Data sparsity is an inherent challenge in the recommender systems, where most
of the data is collected from the implicit feedbacks of users. This causes two
difficulties in designing effective algorithms: first, the majority of users
only have a few interactions with the system and there is no enough data for
learning; second, there are no negative samples in the implicit feedbacks and
it is a common practice to perform negative sampling to generate negative
samples. However, this leads to a consequence that many potential positive
samples are mislabeled as negative ones and data sparsity would exacerbate the
mislabeling problem. To solve these difficulties, we regard the problem of
recommendation on sparse implicit feedbacks as a semi-supervised learning task,
and explore domain adaption to solve it. We transfer the knowledge learned from
dense data to sparse data and we focus on the most challenging case -- there is
no user or item overlap. In this extreme case, aligning embeddings of two
datasets directly is rather sub-optimal since the two latent spaces encode very
different information. As such, we adopt domain-invariant textual features as
the anchor points to align the latent spaces. To align the embeddings, we
extract the textual features for each user and item and feed them into a domain
classifier with the embeddings of users and items. The embeddings are trained
to puzzle the classifier and textual features are fixed as anchor points. By
domain adaptation, the distribution pattern in the source domain is transferred
to the target domain. As the target part can be supervised by domain
adaptation, we abandon negative sampling in target dataset to avoid label
noise. We adopt three pairs of real-world datasets to validate the
effectiveness of our transfer strategy. Results show that our models outperform
existing models significantly.
Related papers
- Downstream-Pretext Domain Knowledge Traceback for Active Learning [138.02530777915362]
We propose a downstream-pretext domain knowledge traceback (DOKT) method that traces the data interactions of downstream knowledge and pre-training guidance.
DOKT consists of a traceback diversity indicator and a domain-based uncertainty estimator.
Experiments conducted on ten datasets show that our model outperforms other state-of-the-art methods.
arXiv Detail & Related papers (2024-07-20T01:34:13Z) - High-order Neighborhoods Know More: HyperGraph Learning Meets Source-free Unsupervised Domain Adaptation [34.08681468394247]
Source-free Unsupervised Domain Adaptation aims to classify target samples by only accessing a pre-trained source model and unlabelled target samples.
Existing methods normally exploit the pair-wise relation among target samples and attempt to discover their correlations by clustering these samples based on semantic features.
We propose a new SFDA method that exploits the high-order neighborhood relation and explicitly takes the domain shift effect into account.
arXiv Detail & Related papers (2024-05-11T05:07:43Z) - Holistic Transfer: Towards Non-Disruptive Fine-Tuning with Partial
Target Data [32.91362206231936]
We propose a learning problem involving adapting a pre-trained source model to the target domain for classifying all classes that appeared in the source data.
This problem is practical, as it is unrealistic for the target end-users to collect data for all classes prior to adaptation.
We present several effective solutions that maintain the accuracy of the missing classes and enhance the overall performance.
arXiv Detail & Related papers (2023-11-02T17:35:16Z) - Divide and Contrast: Source-free Domain Adaptation via Adaptive
Contrastive Learning [122.62311703151215]
Divide and Contrast (DaC) aims to connect the good ends of both worlds while bypassing their limitations.
DaC divides the target data into source-like and target-specific samples, where either group of samples is treated with tailored goals.
We further align the source-like domain with the target-specific samples using a memory bank-based Maximum Mean Discrepancy (MMD) loss to reduce the distribution mismatch.
arXiv Detail & Related papers (2022-11-12T09:21:49Z) - AcroFOD: An Adaptive Method for Cross-domain Few-shot Object Detection [59.10314662986463]
Cross-domain few-shot object detection aims to adapt object detectors in the target domain with a few annotated target data.
The proposed method achieves state-of-the-art performance on multiple benchmarks.
arXiv Detail & Related papers (2022-09-22T10:23:40Z) - Low-confidence Samples Matter for Domain Adaptation [47.552605279925736]
Domain adaptation (DA) aims to transfer knowledge from a label-rich source domain to a related but label-scarce target domain.
We propose a novel contrastive learning method by processing low-confidence samples.
We evaluate the proposed method in both unsupervised and semi-supervised DA settings.
arXiv Detail & Related papers (2022-02-06T15:45:45Z) - Attentive Prototypes for Source-free Unsupervised Domain Adaptive 3D
Object Detection [85.11649974840758]
3D object detection networks tend to be biased towards the data they are trained on.
We propose a single-frame approach for source-free, unsupervised domain adaptation of lidar-based 3D object detectors.
arXiv Detail & Related papers (2021-11-30T18:42:42Z) - Domain Adaptive Semantic Segmentation without Source Data [50.18389578589789]
We investigate domain adaptive semantic segmentation without source data, which assumes that the model is pre-trained on the source domain.
We propose an effective framework for this challenging problem with two components: positive learning and negative learning.
Our framework can be easily implemented and incorporated with other methods to further enhance the performance.
arXiv Detail & Related papers (2021-10-13T04:12:27Z) - Alleviating Semantic-level Shift: A Semi-supervised Domain Adaptation
Method for Semantic Segmentation [97.8552697905657]
A key challenge of this task is how to alleviate the data distribution discrepancy between the source and target domains.
We propose Alleviating Semantic-level Shift (ASS), which can successfully promote the distribution consistency from both global and local views.
We apply our ASS to two domain adaptation tasks, from GTA5 to Cityscapes and from Synthia to Cityscapes.
arXiv Detail & Related papers (2020-04-02T03:25:05Z) - Person Re-identification: Implicitly Defining the Receptive Fields of
Deep Learning Classification Frameworks [5.123298347655088]
This paper describes a solution for implicitly driving the inference of the networks' receptive fields.
We use a segmentation module to distinguish between the foreground (important)/background (irrelevant) parts of each learning instance.
This strategy typically drives the networks to early convergence and appropriate solutions, where the identity and descriptions are not correlated.
arXiv Detail & Related papers (2020-01-30T11:45:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.