Sparks of Artificial General Recommender (AGR): Early Experiments with
ChatGPT
- URL: http://arxiv.org/abs/2305.04518v1
- Date: Mon, 8 May 2023 07:28:16 GMT
- Title: Sparks of Artificial General Recommender (AGR): Early Experiments with
ChatGPT
- Authors: Guo Lin and Yongfeng Zhang
- Abstract summary: An AGR comprises both conversationality and universality to engage in natural dialogues and generate recommendations across various domains.
We propose ten fundamental principles that an AGR should adhere to, each with its corresponding testing protocols.
We assess whether ChatGPT, a sophisticated LLM, can comply with the proposed principles by engaging in recommendation-oriented dialogues with the model while observing its behavior.
- Score: 33.424692414746836
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This study investigates the feasibility of developing an Artificial General
Recommender (AGR), facilitated by recent advancements in Large Language Models
(LLMs). An AGR comprises both conversationality and universality to engage in
natural dialogues and generate recommendations across various domains. We
propose ten fundamental principles that an AGR should adhere to, each with its
corresponding testing protocols. We proceed to assess whether ChatGPT, a
sophisticated LLM, can comply with the proposed principles by engaging in
recommendation-oriented dialogues with the model while observing its behavior.
Our findings demonstrate the potential for ChatGPT to serve as an AGR, though
several limitations and areas for improvement are identified.
Related papers
- Adapting Large Language Models for Multi-Domain Retrieval-Augmented-Generation [59.58987161199141]
Multi-domain applications face challenges like lack of diverse benchmarks and poor out-of-domain generalization.
We introduce a diverse benchmark comprising a variety of question-answering tasks from 8 sources and covering 13 domains.
Our findings highlight key strategies for improving multi-domain RAG robustness.
arXiv Detail & Related papers (2025-04-03T09:03:40Z) - Graph Retrieval-Augmented LLM for Conversational Recommendation Systems [52.35491420330534]
G-CRS (Graph Retrieval-Augmented Large Language Model for Conversational Recommender Systems) is a training-free framework that combines graph retrieval-augmented generation and in-context learning.
G-CRS achieves superior recommendation performance compared to existing methods without requiring task-specific training.
arXiv Detail & Related papers (2025-03-09T03:56:22Z) - CORAL: Benchmarking Multi-turn Conversational Retrieval-Augmentation Generation [68.81271028921647]
We introduce CORAL, a benchmark designed to assess RAG systems in realistic multi-turn conversational settings.
CORAL includes diverse information-seeking conversations automatically derived from Wikipedia.
It supports three core tasks of conversational RAG: passage retrieval, response generation, and citation labeling.
arXiv Detail & Related papers (2024-10-30T15:06:32Z) - R^3AG: First Workshop on Refined and Reliable Retrieval Augmented Generation [30.045100489254327]
Retrieval-augmented generation (RAG) has gained wide attention as the key component to improve generative models with external knowledge augmentation from information retrieval.
This workshop aims to explore in depth how to conduct refined and reliable RAG for downstream AI tasks.
arXiv Detail & Related papers (2024-10-27T21:12:12Z) - Ground Every Sentence: Improving Retrieval-Augmented LLMs with Interleaved Reference-Claim Generation [51.8188846284153]
RAG has been widely adopted to enhance Large Language Models (LLMs)
Attributed Text Generation (ATG) has attracted growing attention, which provides citations to support the model's responses in RAG.
This paper proposes a fine-grained ATG method called ReClaim(Refer & Claim), which alternates the generation of references and answers step by step.
arXiv Detail & Related papers (2024-07-01T20:47:47Z) - Reinforcement of Explainability of ChatGPT Prompts by Embedding Breast Cancer Self-Screening Rules into AI Responses [0.7673339435080445]
The research aims to evaluate ChatGPT's reasoning capabilities, emphasizing its potential to process rules and provide explanations for screening recommendations.
The methodology employs a supervised prompt-engineering approach to enforce detailed explanations for ChatGPT's recommendations.
Findings highlight ChatGPT's promising capacity in processing rules comparable to Expert System Shells.
arXiv Detail & Related papers (2024-04-21T09:20:16Z) - GenRES: Rethinking Evaluation for Generative Relation Extraction in the
Era of Large Language Models [48.56814147033251]
We introduce GenRES for a multi-dimensional assessment in terms of the topic similarity, uniqueness, granularity, factualness, and completeness of the GRE results.
With GenRES, we empirically identified that precision/recall fails to justify the performance of GRE methods.
Next, we conducted a human evaluation of GRE methods that shows GenRES is consistent with human preferences for RE quality.
arXiv Detail & Related papers (2024-02-16T15:01:24Z) - Large Language Models Meet Open-World Intent Discovery and Recognition:
An Evaluation of ChatGPT [37.27411474856601]
Out-of-domain (OOD) intent discovery and generalized intent discovery (GID) aim to extend a closed intent to open-world intent sets.
Previous methods address them by fine-tuning discriminative models.
ChatGPT exhibits consistent advantages under zero-shot settings, but is still at a disadvantage compared to fine-tuned models.
arXiv Detail & Related papers (2023-10-16T08:34:44Z) - Evaluating ChatGPT as a Recommender System: A Rigorous Approach [12.458752059072706]
We propose a robust evaluation pipeline to assess ChatGPT's ability as an RS and post-process ChatGPT recommendations.
We analyze the model's functionality in three settings: the Top-N Recommendation, the cold-start recommendation, and the re-ranking of a list of recommendations.
arXiv Detail & Related papers (2023-09-07T10:13:09Z) - Development of the ChatGPT, Generative Artificial Intelligence and
Natural Large Language Models for Accountable Reporting and Use (CANGARU)
Guidelines [0.33249867230903685]
CANGARU aims to foster a cross-disciplinary global consensus on the ethical use, disclosure, and proper reporting of GAI/GPT/LLM technologies in academia.
The present protocol consists of an ongoing systematic review of GAI/GPT/LLM applications to understand the linked ideas, findings, and reporting standards in scholarly research, and to formulate guidelines for its use and disclosure.
arXiv Detail & Related papers (2023-07-18T05:12:52Z) - Rethinking the Evaluation for Conversational Recommendation in the Era
of Large Language Models [115.7508325840751]
The recent success of large language models (LLMs) has shown great potential to develop more powerful conversational recommender systems (CRSs)
In this paper, we embark on an investigation into the utilization of ChatGPT for conversational recommendation, revealing the inadequacy of the existing evaluation protocol.
We propose an interactive Evaluation approach based on LLMs named iEvaLM that harnesses LLM-based user simulators.
arXiv Detail & Related papers (2023-05-22T15:12:43Z) - Deep Conversational Recommender Systems: A New Frontier for
Goal-Oriented Dialogue Systems [54.06971074217952]
Conversational Recommender System (CRS) learns and models user's preferences through interactive dialogue conversations.
Deep learning approaches are applied to CRS and have produced fruitful results.
arXiv Detail & Related papers (2020-04-28T02:20:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.