Zero-shot Item-based Recommendation via Multi-task Product Knowledge
Graph Pre-Training
- URL: http://arxiv.org/abs/2305.07633v1
- Date: Fri, 12 May 2023 17:38:24 GMT
- Title: Zero-shot Item-based Recommendation via Multi-task Product Knowledge
Graph Pre-Training
- Authors: Ziwei Fan, Zhiwei Liu, Shelby Heinecke, Jianguo Zhang, Huan Wang,
Caiming Xiong, and Philip S. Yu
- Abstract summary: This paper presents a novel paradigm for the Zero-Shot Item-based Recommendation (ZSIR) task.
It pre-trains a model on product knowledge graph (PKG) to refine the item features from PLMs.
We identify three challenges for pre-training PKG, which are multi-type relations in PKG, semantic divergence between item generic information and relations and domain discrepancy from PKG to downstream ZSIR task.
- Score: 106.85813323510783
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Existing recommender systems face difficulties with zero-shot items, i.e.
items that have no historical interactions with users during the training
stage. Though recent works extract universal item representation via
pre-trained language models (PLMs), they ignore the crucial item relationships.
This paper presents a novel paradigm for the Zero-Shot Item-based
Recommendation (ZSIR) task, which pre-trains a model on product knowledge graph
(PKG) to refine the item features from PLMs. We identify three challenges for
pre-training PKG, which are multi-type relations in PKG, semantic divergence
between item generic information and relations and domain discrepancy from PKG
to downstream ZSIR task. We address the challenges by proposing four
pre-training tasks and novel task-oriented adaptation (ToA) layers. Moreover,
this paper discusses how to fine-tune the model on new recommendation task such
that the ToA layers are adapted to ZSIR task. Comprehensive experiments on 18
markets dataset are conducted to verify the effectiveness of the proposed model
in both knowledge prediction and ZSIR task.
Related papers
- P-RAG: Progressive Retrieval Augmented Generation For Planning on Embodied Everyday Task [94.08478298711789]
Embodied Everyday Task is a popular task in the embodied AI community.
Natural language instructions often lack explicit task planning.
Extensive training is required to equip models with knowledge of the task environment.
arXiv Detail & Related papers (2024-09-17T15:29:34Z) - On Task-personalized Multimodal Few-shot Learning for Visually-rich
Document Entity Retrieval [59.25292920967197]
Few-shot document entity retrieval (VDER) is an important topic in industrial NLP applications.
FewVEX is a new dataset to boost future research in the field of entity-level few-shot VDER.
We present a task-aware meta-learning based framework, with a central focus on achieving effective task personalization.
arXiv Detail & Related papers (2023-11-01T17:51:43Z) - Analysis of the Reasoning with Redundant Information Provided Ability of
Large Language Models [0.0]
Large Language Models (LLMs) have demonstrated impressive capabilities across a range of natural language processing tasks.
To address this gap, a new form of Question-Answering (QA) task, termed Reasoning with Redundant Information Provided (RRIP), is introduced.
This study evaluates two popular LLMs, LlaMA2-13B-chat and generative pre-trained transformer 3.5 (GPT-3.5), contrasting their performance on traditional QA tasks against RRIP tasks.
arXiv Detail & Related papers (2023-10-06T06:20:06Z) - Information Association for Language Model Updating by Mitigating
LM-Logical Discrepancy [68.31760483418901]
Large Language Models(LLMs) struggle with providing current information due to the outdated pre-training data.
Existing methods for updating LLMs, such as knowledge editing and continual fine-tuning, have significant drawbacks in generalizability of new information.
We identify the core challenge behind these drawbacks: the LM-logical discrepancy featuring the difference between language modeling probabilities and logical probabilities.
arXiv Detail & Related papers (2023-05-29T19:48:37Z) - Leveraging Pre-Trained Language Models to Streamline Natural Language
Interaction for Self-Tracking [25.28975864365579]
We propose a novel NLP task for self-tracking that extracts close- and open-ended information from a retrospective activity log.
The framework augments the prompt using synthetic samples to transform the task into 10-shot learning, to address a cold-start problem in bootstrapping a new tracking topic.
arXiv Detail & Related papers (2022-05-31T01:58:04Z) - CINS: Comprehensive Instruction for Few-shot Learning in Task-oriented
Dialog Systems [56.302581679816775]
This paper proposes Comprehensive Instruction (CINS) that exploits PLMs with task-specific instructions.
We design a schema (definition, constraint, prompt) of instructions and their customized realizations for three important downstream tasks in ToD.
Experiments are conducted on these ToD tasks in realistic few-shot learning scenarios with small validation data.
arXiv Detail & Related papers (2021-09-10T03:23:06Z) - Weighted Training for Cross-Task Learning [71.94908559469475]
We introduce Target-Aware Weighted Training (TAWT), a weighted training algorithm for cross-task learning.
We show that TAWT is easy to implement, is computationally efficient, requires little hyper parameter tuning, and enjoys non-asymptotic learning-theoretic guarantees.
As a byproduct, the proposed representation-based task distance allows one to reason in a theoretically principled way about several critical aspects of cross-task learning.
arXiv Detail & Related papers (2021-05-28T20:27:02Z) - Curriculum-Meta Learning for Order-Robust Continual Relation Extraction [12.494209368988253]
We propose a novel curriculum-meta learning method to tackle the challenges of continual relation extraction.
We combine meta learning and curriculum learning to quickly adapt model parameters to a new task.
We present novel difficulty-based metrics to quantitatively measure the extent of order-sensitivity of a given model.
arXiv Detail & Related papers (2021-01-06T08:52:34Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.