Unified Pretraining for Recommendation via Task Hypergraphs
- URL: http://arxiv.org/abs/2310.13286v1
- Date: Fri, 20 Oct 2023 05:33:21 GMT
- Title: Unified Pretraining for Recommendation via Task Hypergraphs
- Authors: Mingdai Yang, Zhiwei Liu, Liangwei Yang, Xiaolong Liu, Chen Wang, Hao
Peng, Philip S. Yu
- Abstract summary: We propose a novel multitask pretraining framework named Unified Pretraining for Recommendation via Task Hypergraphs.
For a unified learning pattern to handle diverse requirements and nuances of various pretext tasks, we design task hypergraphs to generalize pretext tasks to hyperedge prediction.
A novel transitional attention layer is devised to discriminatively learn the relevance between each pretext task and recommendation.
- Score: 55.98773629788986
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Although pretraining has garnered significant attention and popularity in
recent years, its application in graph-based recommender systems is relatively
limited. It is challenging to exploit prior knowledge by pretraining in widely
used ID-dependent datasets. On one hand, user-item interaction history in one
dataset can hardly be transferred to other datasets through pretraining, where
IDs are different. On the other hand, pretraining and finetuning on the same
dataset leads to a high risk of overfitting. In this paper, we propose a novel
multitask pretraining framework named Unified Pretraining for Recommendation
via Task Hypergraphs. For a unified learning pattern to handle diverse
requirements and nuances of various pretext tasks, we design task hypergraphs
to generalize pretext tasks to hyperedge prediction. A novel transitional
attention layer is devised to discriminatively learn the relevance between each
pretext task and recommendation. Experimental results on three benchmark
datasets verify the superiority of UPRTH. Additional detailed investigations
are conducted to demonstrate the effectiveness of the proposed framework.
Related papers
- MultiGPrompt for Multi-Task Pre-Training and Prompting on Graphs [33.2696184519275]
MultiGPrompt is a novel multi-task pre-training and prompting framework for graph representation learning.
We propose a dual-prompt mechanism consisting of composed and open prompts to leverage task-specific and global pre-training knowledge.
arXiv Detail & Related papers (2023-11-28T02:36:53Z) - A Supervised Contrastive Learning Pretrain-Finetune Approach for Time
Series [15.218841180577135]
We introduce a novel pretraining procedure that leverages supervised contrastive learning to distinguish features within each pretraining dataset.
We then propose a fine-tuning procedure designed to enhance the accurate prediction of the target data by aligning it more closely with the learned dynamics of the pretraining datasets.
arXiv Detail & Related papers (2023-11-21T02:06:52Z) - ULTRA-DP: Unifying Graph Pre-training with Multi-task Graph Dual Prompt [67.8934749027315]
We propose a unified framework for graph hybrid pre-training which injects the task identification and position identification into GNNs.
We also propose a novel pre-training paradigm based on a group of $k$-nearest neighbors.
arXiv Detail & Related papers (2023-10-23T12:11:13Z) - Socratic Pretraining: Question-Driven Pretraining for Controllable
Summarization [89.04537372465612]
Socratic pretraining is a question-driven, unsupervised pretraining objective designed to improve controllability in summarization tasks.
Our results show that Socratic pretraining cuts task-specific labeled data requirements in half.
arXiv Detail & Related papers (2022-12-20T17:27:10Z) - Improved Fine-tuning by Leveraging Pre-training Data: Theory and
Practice [52.11183787786718]
Fine-tuning a pre-trained model on the target data is widely used in many deep learning applications.
Recent studies have empirically shown that training from scratch has the final performance that is no worse than this pre-training strategy.
We propose a novel selection strategy to select a subset from pre-training data to help improve the generalization on the target task.
arXiv Detail & Related papers (2021-11-24T06:18:32Z) - KGPT: Knowledge-Grounded Pre-Training for Data-to-Text Generation [100.79870384880333]
We propose a knowledge-grounded pre-training (KGPT) to generate knowledge-enriched text.
We adopt three settings, namely fully-supervised, zero-shot, few-shot to evaluate its effectiveness.
Under zero-shot setting, our model achieves over 30 ROUGE-L on WebNLG while all other baselines fail.
arXiv Detail & Related papers (2020-10-05T19:59:05Z) - Omni-supervised Facial Expression Recognition via Distilled Data [120.11782405714234]
We propose omni-supervised learning to exploit reliable samples in a large amount of unlabeled data for network training.
We experimentally verify that the new dataset can significantly improve the ability of the learned FER model.
To tackle this, we propose to apply a dataset distillation strategy to compress the created dataset into several informative class-wise images.
arXiv Detail & Related papers (2020-05-18T09:36:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.