A General Offline Reinforcement Learning Framework for Interactive
Recommendation
- URL: http://arxiv.org/abs/2310.00678v1
- Date: Sun, 1 Oct 2023 14:09:21 GMT
- Title: A General Offline Reinforcement Learning Framework for Interactive
Recommendation
- Authors: Teng Xiao, Donglin Wang
- Abstract summary: We first introduce a probabilistic generative model for interactive recommendation, and then propose an effective inference algorithm for discrete and policy learning based on logged feedbacks.
We conduct extensive experiments on two public real-world datasets, demonstrating that the proposed methods can achieve superior performance over existing supervised learning and reinforcement learning methods for recommendation.
- Score: 43.47849328010646
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper studies the problem of learning interactive recommender systems
from logged feedbacks without any exploration in online environments. We
address the problem by proposing a general offline reinforcement learning
framework for recommendation, which enables maximizing cumulative user rewards
without online exploration. Specifically, we first introduce a probabilistic
generative model for interactive recommendation, and then propose an effective
inference algorithm for discrete and stochastic policy learning based on logged
feedbacks. In order to perform offline learning more effectively, we propose
five approaches to minimize the distribution mismatch between the logging
policy and recommendation policy: support constraints, supervised
regularization, policy constraints, dual constraints and reward extrapolation.
We conduct extensive experiments on two public real-world datasets,
demonstrating that the proposed methods can achieve superior performance over
existing supervised learning and reinforcement learning methods for
recommendation.
Related papers
- Adversarial Batch Inverse Reinforcement Learning: Learn to Reward from
Imperfect Demonstration for Interactive Recommendation [23.048841953423846]
We focus on the problem of learning to reward, which is fundamental to reinforcement learning.
Previous approaches either introduce additional procedures for learning to reward, thereby increasing the complexity of optimization.
We propose a novel batch inverse reinforcement learning paradigm that achieves the desired properties.
arXiv Detail & Related papers (2023-10-30T13:43:20Z) - AdaptiveRec: Adaptively Construct Pairs for Contrastive Learning in
Sequential Recommendation [0.7883397954991659]
This paper presents a solution to the challenges faced by contrastive learning in sequential recommendation systems.
It addresses the issue of false negative, which limits the effectiveness of recommendation algorithms.
arXiv Detail & Related papers (2023-07-07T06:48:58Z) - A Regularized Implicit Policy for Offline Reinforcement Learning [54.7427227775581]
offline reinforcement learning enables learning from a fixed dataset, without further interactions with the environment.
We propose a framework that supports learning a flexible yet well-regularized fully-implicit policy.
Experiments and ablation study on the D4RL dataset validate our framework and the effectiveness of our algorithmic designs.
arXiv Detail & Related papers (2022-02-19T20:22:04Z) - Off-policy Reinforcement Learning with Optimistic Exploration and
Distribution Correction [73.77593805292194]
We train a separate exploration policy to maximize an approximate upper confidence bound of the critics in an off-policy actor-critic framework.
To mitigate the off-policy-ness, we adapt the recently introduced DICE framework to learn a distribution correction ratio for off-policy actor-critic training.
arXiv Detail & Related papers (2021-10-22T22:07:51Z) - Offline Meta-level Model-based Reinforcement Learning Approach for
Cold-Start Recommendation [27.17948754183511]
Reinforcement learning has shown great promise in optimizing long-term user interest in recommender systems.
Existing RL-based recommendation methods need a large number of interactions for each user to learn a robust recommendation policy.
We propose a meta-level model-based reinforcement learning approach for fast user adaptation.
arXiv Detail & Related papers (2020-12-04T08:58:35Z) - Adversarial Counterfactual Learning and Evaluation for Recommender
System [33.44276155380476]
We show in theory that applying supervised learning to detect user preferences may end up with inconsistent results in the absence of exposure information.
We propose a principled solution by introducing a minimax empirical risk formulation.
arXiv Detail & Related papers (2020-11-08T00:40:51Z) - Generative Inverse Deep Reinforcement Learning for Online Recommendation [62.09946317831129]
We propose a novel inverse reinforcement learning approach, namely InvRec, for online recommendation.
InvRec extracts the reward function from user's behaviors automatically, for online recommendation.
arXiv Detail & Related papers (2020-11-04T12:12:25Z) - Self-Supervised Reinforcement Learning for Recommender Systems [77.38665506495553]
We propose self-supervised reinforcement learning for sequential recommendation tasks.
Our approach augments standard recommendation models with two output layers: one for self-supervised learning and the other for RL.
Based on such an approach, we propose two frameworks namely Self-Supervised Q-learning(SQN) and Self-Supervised Actor-Critic(SAC)
arXiv Detail & Related papers (2020-06-10T11:18:57Z) - Reward-Conditioned Policies [100.64167842905069]
imitation learning requires near-optimal expert data.
Can we learn effective policies via supervised learning without demonstrations?
We show how such an approach can be derived as a principled method for policy search.
arXiv Detail & Related papers (2019-12-31T18:07:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.