MGFRec: Towards Reinforced Reasoning Recommendation with Multiple Groundings and Feedback
- URL: http://arxiv.org/abs/2510.22888v1
- Date: Mon, 27 Oct 2025 00:41:07 GMT
- Title: MGFRec: Towards Reinforced Reasoning Recommendation with Multiple Groundings and Feedback
- Authors: Shihao Cai, Chongming Gao, Haoyan Liu, Wentao Shi, Jianshan Sun, Ruiming Tang, Fuli Feng,
- Abstract summary: We propose performing multiple rounds of grounding during inference to help the LLM better understand the actual item space.<n> Comprehensive experiments conducted on three Amazon review datasets demonstrate the effectiveness of incorporating multiple groundings and feedback.
- Score: 62.59727494001646
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The powerful reasoning and generative capabilities of large language models (LLMs) have inspired researchers to apply them to reasoning-based recommendation tasks, which require in-depth reasoning about user interests and the generation of recommended items. However, previous reasoning-based recommendation methods have typically performed inference within the language space alone, without incorporating the actual item space. This has led to over-interpreting user interests and deviating from real items. Towards this research gap, we propose performing multiple rounds of grounding during inference to help the LLM better understand the actual item space, which could ensure that its reasoning remains aligned with real items. Furthermore, we introduce a user agent that provides feedback during each grounding step, enabling the LLM to better recognize and adapt to user interests. Comprehensive experiments conducted on three Amazon review datasets demonstrate the effectiveness of incorporating multiple groundings and feedback. These findings underscore the critical importance of reasoning within the actual item space, rather than being confined to the language space, for recommendation tasks.
Related papers
- Enhancing LLM-based Recommendation with Preference Hint Discovery from Knowledge Graph [26.065016082075264]
We propose a preference hint discovery model based on the interaction-integrated knowledge graph.<n>We develop an instance-wise dual-attention mechanism to quantify the preference credibility of candidate attributes.
arXiv Detail & Related papers (2026-01-26T03:20:42Z) - OneRec-Think: In-Text Reasoning for Generative Recommendation [55.53292983432484]
OneRec-Think is a unified framework that seamlessly integrates dialogue, reasoning, and personalized recommendation.<n>Our proposed "Think-Ahead" architecture enables effective industrial deployment on Kuaishou, achieving a 0.159% gain in APP Stay Time.
arXiv Detail & Related papers (2025-10-13T17:20:13Z) - From What to Why: Thought-Space Recommendation with Small Language Models [2.134948383299948]
Large Language Models (LLMs) have advanced recommendation capabilities through enhanced reasoning, but pose significant challenges for real-world deployment.<n>Small Language Models (SLMs) offer an efficient alternative, their reasoning capabilities for recommendation remain underexplored.<n>We propose PULSE (Preference Understanding by Latent Semantic Embeddings), a framework that treats SLM-generated rationales as director learning signals.
arXiv Detail & Related papers (2025-10-08T11:22:26Z) - Learning to Shop Like Humans: A Review-driven Retrieval-Augmented Recommendation Framework with LLMs [30.748667156183004]
RevBrowse is a review-driven recommendation framework inspired by the "browse-then-decide" decision process.<n>RevBrowse integrates user reviews into the LLM-based reranking process to enhance its ability to distinguish between candidate items.<n>PrefRAG is a retrieval-augmented module that disentangles user and item representations into structured forms.
arXiv Detail & Related papers (2025-08-31T04:37:43Z) - Evaluating Large Language Models as Expert Annotators [17.06186816803593]
This paper investigates whether top-performing language models can serve as direct alternatives to human expert annotators.<n>We evaluate both individual LLMs and multi-agent approaches across three highly specialized domains: finance, biomedicine, and law.<n>Our empirical results reveal that individual LLMs equipped with inference-time techniques show only marginal or even negative performance gains.
arXiv Detail & Related papers (2025-08-11T10:19:10Z) - Review-driven Personalized Preference Reasoning with Large Language Models for Recommendation [21.769969074938142]
EXP3RT is a novel LLM-based recommender designed to leverage rich preference information contained in user and item reviews.<n>It generates detailed step-by-step reasoning followed by predicted rating.<n>Experiments show that EXP3RT outperforms existing methods on both rating prediction and candidate item reranking for top-k recommendation.
arXiv Detail & Related papers (2024-08-12T16:39:03Z) - Language Representations Can be What Recommenders Need: Findings and Potentials [57.90679739598295]
We show that item representations, when linearly mapped from advanced LM representations, yield superior recommendation performance.<n>This outcome suggests the possible homomorphism between the advanced language representation space and an effective item representation space for recommendation.<n>Our findings highlight the connection between language modeling and behavior modeling, which can inspire both natural language processing and recommender system communities.
arXiv Detail & Related papers (2024-07-07T17:05:24Z) - LEARN: Knowledge Adaptation from Large Language Model to Recommendation for Practical Industrial Application [54.984348122105516]
Llm-driven knowlEdge Adaptive RecommeNdation (LEARN) framework synergizes open-world knowledge with collaborative knowledge.<n>We propose an Llm-driven knowlEdge Adaptive RecommeNdation (LEARN) framework that synergizes open-world knowledge with collaborative knowledge.
arXiv Detail & Related papers (2024-05-07T04:00:30Z) - Sample Efficient Preference Alignment in LLMs via Active Exploration [63.84454768573154]
We take advantage of the fact that one can often choose contexts at which to obtain human feedback to most efficiently identify a good policy.<n>We propose an active exploration algorithm to efficiently select the data and provide theoretical proof that it has a worst-case regret bound.<n>Our method outperforms the baselines with limited samples of human preferences on several language models and four real-world datasets.
arXiv Detail & Related papers (2023-12-01T00:54:02Z) - Rethinking the Evaluation for Conversational Recommendation in the Era
of Large Language Models [115.7508325840751]
The recent success of large language models (LLMs) has shown great potential to develop more powerful conversational recommender systems (CRSs)
In this paper, we embark on an investigation into the utilization of ChatGPT for conversational recommendation, revealing the inadequacy of the existing evaluation protocol.
We propose an interactive Evaluation approach based on LLMs named iEvaLM that harnesses LLM-based user simulators.
arXiv Detail & Related papers (2023-05-22T15:12:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.