GPT4Rec: A Generative Framework for Personalized Recommendation and User
Interests Interpretation
- URL: http://arxiv.org/abs/2304.03879v1
- Date: Sat, 8 Apr 2023 00:30:08 GMT
- Title: GPT4Rec: A Generative Framework for Personalized Recommendation and User
Interests Interpretation
- Authors: Jinming Li, Wentao Zhang, Tian Wang, Guanglei Xiong, Alan Lu, Gerard
Medioni
- Abstract summary: GPT4Rec is a novel and flexible generative framework inspired by search engines.
It first generates hypothetical "search queries" given item titles in a user's history, and then retrieves items for recommendation by searching these queries.
Our framework outperforms state-of-the-art methods by $75.7%$ and $22.2%$ in Recall@K on two public datasets.
- Score: 8.293646972329581
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recent advancements in Natural Language Processing (NLP) have led to the
development of NLP-based recommender systems that have shown superior
performance. However, current models commonly treat items as mere IDs and adopt
discriminative modeling, resulting in limitations of (1) fully leveraging the
content information of items and the language modeling capabilities of NLP
models; (2) interpreting user interests to improve relevance and diversity; and
(3) adapting practical circumstances such as growing item inventories. To
address these limitations, we present GPT4Rec, a novel and flexible generative
framework inspired by search engines. It first generates hypothetical "search
queries" given item titles in a user's history, and then retrieves items for
recommendation by searching these queries. The framework overcomes previous
limitations by learning both user and item embeddings in the language space. To
well-capture user interests with different aspects and granularity for
improving relevance and diversity, we propose a multi-query generation
technique with beam search. The generated queries naturally serve as
interpretable representations of user interests and can be searched to
recommend cold-start items. With GPT-2 language model and BM25 search engine,
our framework outperforms state-of-the-art methods by $75.7\%$ and $22.2\%$ in
Recall@K on two public datasets. Experiments further revealed that multi-query
generation with beam search improves both the diversity of retrieved items and
the coverage of a user's multi-interests. The adaptiveness and interpretability
of generated queries are discussed with qualitative case studies.
Related papers
- Think-then-Act: A Dual-Angle Evaluated Retrieval-Augmented Generation [3.2134014920850364]
Large language models (LLMs) often face challenges such as temporal misalignment and generating hallucinatory content.
We propose a dual-angle evaluated retrieval-augmented generation framework textitThink-then-Act'
arXiv Detail & Related papers (2024-06-18T20:51:34Z) - Optimized Feature Generation for Tabular Data via LLMs with Decision Tree Reasoning [53.241569810013836]
We propose a new framework based on large language models (LLMs) and decision Tree reasoning (OCTree)
Our key idea is to leverage LLMs' reasoning capabilities to find good feature generation rules without manually specifying the search space.
Our empirical results demonstrate that this simple framework consistently enhances the performance of various prediction models.
arXiv Detail & Related papers (2024-06-12T08:31:34Z) - A Survey of Generative Search and Recommendation in the Era of Large Language Models [125.26354486027408]
generative search (retrieval) and recommendation aims to address the matching problem in a generative manner.
Superintelligent generative large language models have sparked a new paradigm in search and recommendation.
arXiv Detail & Related papers (2024-04-25T17:58:17Z) - Leveraging Large Language Models for Multimodal Search [0.6249768559720121]
This paper introduces a novel multimodal search model that achieves a new performance milestone on the Fashion200K dataset.
We also propose a novel search interface integrating Large Language Models (LLMs) to facilitate natural language interaction.
arXiv Detail & Related papers (2024-04-24T10:30:42Z) - Generative Multi-Modal Knowledge Retrieval with Large Language Models [75.70313858231833]
We propose an innovative end-to-end generative framework for multi-modal knowledge retrieval.
Our framework takes advantage of the fact that large language models (LLMs) can effectively serve as virtual knowledge bases.
We demonstrate significant improvements ranging from 3.0% to 14.6% across all evaluation metrics when compared to strong baselines.
arXiv Detail & Related papers (2024-01-16T08:44:29Z) - Ada-Retrieval: An Adaptive Multi-Round Retrieval Paradigm for Sequential
Recommendations [50.03560306423678]
We propose Ada-Retrieval, an adaptive multi-round retrieval paradigm for recommender systems.
Ada-Retrieval iteratively refines user representations to better capture potential candidates in the full item space.
arXiv Detail & Related papers (2024-01-12T15:26:40Z) - Enhancing Retrieval-Augmented Large Language Models with Iterative
Retrieval-Generation Synergy [164.83371924650294]
We show that strong performance can be achieved by a method we call Iter-RetGen, which synergizes retrieval and generation in an iterative manner.
A model output shows what might be needed to finish a task, and thus provides an informative context for retrieving more relevant knowledge.
Iter-RetGen processes all retrieved knowledge as a whole and largely preserves the flexibility in generation without structural constraints.
arXiv Detail & Related papers (2023-05-24T16:17:36Z) - Recommender Systems with Generative Retrieval [58.454606442670034]
We propose a novel generative retrieval approach, where the retrieval model autoregressively decodes the identifiers of the target candidates.
To that end, we create semantically meaningful of codewords to serve as a Semantic ID for each item.
We show that recommender systems trained with the proposed paradigm significantly outperform the current SOTA models on various datasets.
arXiv Detail & Related papers (2023-05-08T21:48:17Z) - Empirical Evaluation of ChatGPT on Requirements Information Retrieval
Under Zero-Shot Setting [12.733403458944972]
We empirically evaluate ChatGPT's performance on requirements information retrieval tasks.
Under zero-shot setting, evaluation results reveal ChatGPT's promising ability to retrieve requirements relevant information.
arXiv Detail & Related papers (2023-04-25T04:09:45Z) - GQE-PRF: Generative Query Expansion with Pseudo-Relevance Feedback [8.142861977776256]
We propose a novel approach which effectively integrates text generation models into PRF-based query expansion.
Our approach generates augmented query terms via neural text generation models conditioned on both the initial query and pseudo-relevance feedback.
We evaluate the performance of our approach on information retrieval tasks using two benchmark datasets.
arXiv Detail & Related papers (2021-08-13T01:09:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.