Effective Two-Stage Knowledge Transfer for Multi-Entity Cross-Domain
Recommendation
- URL: http://arxiv.org/abs/2402.19101v1
- Date: Thu, 29 Feb 2024 12:29:58 GMT
- Title: Effective Two-Stage Knowledge Transfer for Multi-Entity Cross-Domain
Recommendation
- Authors: Jianyu Guan, Zongming Yin, Tianyi Zhang, Leihui Chen, Yin Zhang, Fei
Huang, Jufeng Chen, Shuguang Han
- Abstract summary: We propose a pre-training & fine-tuning based Multi-entity Knowledge Transfer framework called MKT.
M MKT utilizes a multi-entity pre-training module to extract transferable knowledge across different entities.
In the end, the extracted common knowledge is adopted for target entity model training.
- Score: 27.14804652946457
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, the recommendation content on e-commerce platforms has
become increasingly rich -- a single user feed may contain multiple entities,
such as selling products, short videos, and content posts. To deal with the
multi-entity recommendation problem, an intuitive solution is to adopt the
shared-network-based architecture for joint training. The idea is to transfer
the extracted knowledge from one type of entity (source entity) to another
(target entity). However, different from the conventional same-entity
cross-domain recommendation, multi-entity knowledge transfer encounters several
important issues: (1) data distributions of the source entity and target entity
are naturally different, making the shared-network-based joint training
susceptible to the negative transfer issue, (2) more importantly, the
corresponding feature schema of each entity is not exactly aligned (e.g., price
is an essential feature for selling product while missing for content posts),
making the existing methods no longer appropriate. Recent researchers have also
experimented with the pre-training and fine-tuning paradigm. Again, they only
consider the scenarios with the same entity type and feature systems, which is
inappropriate in our case. To this end, we design a pre-training & fine-tuning
based Multi-entity Knowledge Transfer framework called MKT. MKT utilizes a
multi-entity pre-training module to extract transferable knowledge across
different entities. In particular, a feature alignment module is first applied
to scale and align different feature schemas. Afterward, a couple of knowledge
extractors are employed to extract the common and entity-specific knowledge. In
the end, the extracted common knowledge is adopted for target entity model
training. Through extensive offline and online experiments, we demonstrated the
superiority of MKT over multiple State-Of-The-Art methods.
Related papers
- Two Heads Are Better Than One: Integrating Knowledge from Knowledge
Graphs and Large Language Models for Entity Alignment [31.70064035432789]
We propose a Large Language Model-enhanced Entity Alignment framework (LLMEA)
LLMEA identifies candidate alignments for a given entity by considering both embedding similarities between entities across Knowledge Graphs and edit distances to a virtual equivalent entity.
Experiments conducted on three public datasets reveal that LLMEA surpasses leading baseline models.
arXiv Detail & Related papers (2024-01-30T12:41:04Z) - MESED: A Multi-modal Entity Set Expansion Dataset with Fine-grained
Semantic Classes and Hard Negative Entities [25.059177235004952]
We propose Multi-modal Entity Set Expansion (MESE), where models integrate information from multiple modalities to represent entities.
A powerful multi-modal model MultiExpan is proposed which is pre-trained on four multimodal pre-training tasks.
The MESED dataset is the first multi-modal dataset for ESE with large-scale and elaborate manual calibration.
arXiv Detail & Related papers (2023-07-27T14:09:59Z) - Towards Better Entity Linking with Multi-View Enhanced Distillation [30.554387215553238]
This paper proposes a Multi-View Enhanced Distillation (MVD) framework for entity linking.
MVD can effectively transfer knowledge of multiple fine-grained and mention-relevant parts within entities from cross-encoders to dual-encoders.
Experiments show our method achieves state-of-the-art performance on several entity linking benchmarks.
arXiv Detail & Related papers (2023-05-27T05:15:28Z) - Nested Named Entity Recognition from Medical Texts: An Adaptive Shared
Network Architecture with Attentive CRF [53.55504611255664]
We propose a novel method, referred to as ASAC, to solve the dilemma caused by the nested phenomenon.
The proposed method contains two key modules: the adaptive shared (AS) part and the attentive conditional random field (ACRF) module.
Our model could learn better entity representations by capturing the implicit distinctions and relationships between different categories of entities.
arXiv Detail & Related papers (2022-11-09T09:23:56Z) - Meta Knowledge Condensation for Federated Learning [65.20774786251683]
Existing federated learning paradigms usually extensively exchange distributed models at a central solver to achieve a more powerful model.
This would incur severe communication burden between a server and multiple clients especially when data distributions are heterogeneous.
Unlike existing paradigms, we introduce an alternative perspective to significantly decrease the communication cost in federate learning.
arXiv Detail & Related papers (2022-09-29T15:07:37Z) - Multi-modal Contrastive Representation Learning for Entity Alignment [57.92705405276161]
Multi-modal entity alignment aims to identify equivalent entities between two different multi-modal knowledge graphs.
We propose MCLEA, a Multi-modal Contrastive Learning based Entity Alignment model.
In particular, MCLEA firstly learns multiple individual representations from multiple modalities, and then performs contrastive learning to jointly model intra-modal and inter-modal interactions.
arXiv Detail & Related papers (2022-09-02T08:59:57Z) - Entity-Graph Enhanced Cross-Modal Pretraining for Instance-level Product
Retrieval [152.3504607706575]
This research aims to conduct weakly-supervised multi-modal instance-level product retrieval for fine-grained product categories.
We first contribute the Product1M datasets, and define two real practical instance-level retrieval tasks.
We exploit to train a more effective cross-modal model which is adaptively capable of incorporating key concept information from the multi-modal data.
arXiv Detail & Related papers (2022-06-17T15:40:45Z) - Deep Transfer Learning for Multi-source Entity Linkage via Domain
Adaptation [63.24594955429465]
Multi-source entity linkage is critical in high-impact applications such as data cleaning and user stitching.
AdaMEL is a deep transfer learning framework that learns generic high-level knowledge to perform multi-source entity linkage.
Our framework achieves state-of-the-art results with 8.21% improvement on average over methods based on supervised learning.
arXiv Detail & Related papers (2021-10-27T15:20:41Z) - Learning to Combine: Knowledge Aggregation for Multi-Source Domain
Adaptation [56.694330303488435]
We propose a Learning to Combine for Multi-Source Domain Adaptation (LtC-MSDA) framework.
In the nutshell, a knowledge graph is constructed on the prototypes of various domains to realize the information propagation among semantically adjacent representations.
Our approach outperforms existing methods with a remarkable margin.
arXiv Detail & Related papers (2020-07-17T07:52:44Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.