Cluster Based Deep Contextual Reinforcement Learning for top-k
Recommendations
- URL: http://arxiv.org/abs/2012.02291v1
- Date: Sun, 29 Nov 2020 20:24:39 GMT
- Title: Cluster Based Deep Contextual Reinforcement Learning for top-k
Recommendations
- Authors: Anubha Kabra, Anu Agarwal, Anil Singh Parihar
- Abstract summary: We propose a novel method for generating top-k recommendations by creating an ensemble of clustering with reinforcement learning.
We have incorporated DB Scan clustering to tackle vast item space, hence in-creasing the efficiency multi-fold.
With partial updates and batch updates, the model learns user patterns continuously.
- Score: 2.8207195763355704
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Rapid advancements in the E-commerce sector over the last few decades have
led to an imminent need for personalised, efficient and dynamic recommendation
systems. To sufficiently cater to this need, we propose a novel method for
generating top-k recommendations by creating an ensemble of clustering with
reinforcement learning. We have incorporated DB Scan clustering to tackle vast
item space, hence in-creasing the efficiency multi-fold. Moreover, by using
deep contextual reinforcement learning, our proposed work leverages the user
features to its full potential. With partial updates and batch updates, the
model learns user patterns continuously. The Duelling Bandit based exploration
provides robust exploration as compared to the state-of-art strategies due to
its adaptive nature. Detailed experiments conducted on a public dataset verify
our claims about the efficiency of our technique as com-pared to existing
techniques.
Related papers
- Knowledge Adaptation from Large Language Model to Recommendation for Practical Industrial Application [54.984348122105516]
Large Language Models (LLMs) pretrained on massive text corpus presents a promising avenue for enhancing recommender systems.
We propose an Llm-driven knowlEdge Adaptive RecommeNdation (LEARN) framework that synergizes open-world knowledge with collaborative knowledge.
arXiv Detail & Related papers (2024-05-07T04:00:30Z) - Embedding in Recommender Systems: A Survey [67.67966158305603]
A crucial aspect is embedding techniques that covert the high-dimensional discrete features, such as user and item IDs, into low-dimensional continuous vectors.
Applying embedding techniques captures complex entity relationships and has spurred substantial research.
This survey covers embedding methods like collaborative filtering, self-supervised learning, and graph-based techniques.
arXiv Detail & Related papers (2023-10-28T06:31:06Z) - Hierarchical Reinforcement Learning for Modeling User Novelty-Seeking
Intent in Recommender Systems [26.519571240032967]
We propose a novel hierarchical reinforcement learning-based method to model the hierarchical user novelty-seeking intent.
We further incorporate diversity and novelty-related measurement in the reward function of the hierarchical RL (HRL) agent to encourage user exploration.
arXiv Detail & Related papers (2023-06-02T12:02:23Z) - What Matters in Reinforcement Learning for Tractography [12.940129711489005]
Deep reinforcement learning (RL) has been proposed to learn the tractography procedure and train agents to reconstruct the structure of the white matter without manually curated reference streamlines.
We thoroughly explore the different components of the proposed framework, such as the choice of the RL algorithm, seeding strategy, the input signal and reward function, and shed light on their impact.
We propose a series of recommendations concerning the choice of RL algorithm, the input to the agents, the reward function and more to help future work using reinforcement learning for tractography.
arXiv Detail & Related papers (2023-05-15T22:01:48Z) - SURF: Semi-supervised Reward Learning with Data Augmentation for
Feedback-efficient Preference-based Reinforcement Learning [168.89470249446023]
We present SURF, a semi-supervised reward learning framework that utilizes a large amount of unlabeled samples with data augmentation.
In order to leverage unlabeled samples for reward learning, we infer pseudo-labels of the unlabeled samples based on the confidence of the preference predictor.
Our experiments demonstrate that our approach significantly improves the feedback-efficiency of the preference-based method on a variety of locomotion and robotic manipulation tasks.
arXiv Detail & Related papers (2022-03-18T16:50:38Z) - A Deep Reinforcement Learning Approach for Composing Moving IoT Services [0.12891210250935145]
We introduce a moving crowdsourced service model which is modelled as a moving region.
We propose a deep reinforcement learning-based composition approach to select and compose moving IoT services.
The experiments on two real-world datasets verify the effectiveness and efficiency of the deep reinforcement learning-based approach.
arXiv Detail & Related papers (2021-11-06T22:02:31Z) - D2RLIR : an improved and diversified ranking function in interactive
recommendation systems based on deep reinforcement learning [0.3058685580689604]
This paper proposes a deep reinforcement learning based recommendation system by utilizing Actor-Critic architecture.
The proposed model is able to generate a diverse while relevance recommendation list based on the user's preferences.
arXiv Detail & Related papers (2021-10-28T13:11:29Z) - Generative Inverse Deep Reinforcement Learning for Online Recommendation [62.09946317831129]
We propose a novel inverse reinforcement learning approach, namely InvRec, for online recommendation.
InvRec extracts the reward function from user's behaviors automatically, for online recommendation.
arXiv Detail & Related papers (2020-11-04T12:12:25Z) - Knowledge-guided Deep Reinforcement Learning for Interactive
Recommendation [49.32287384774351]
Interactive recommendation aims to learn from dynamic interactions between items and users to achieve responsiveness and accuracy.
We propose Knowledge-Guided deep Reinforcement learning to harness the advantages of both reinforcement learning and knowledge graphs for interactive recommendation.
arXiv Detail & Related papers (2020-04-17T05:26:47Z) - Learning From Multiple Experts: Self-paced Knowledge Distillation for
Long-tailed Classification [106.08067870620218]
We propose a self-paced knowledge distillation framework, termed Learning From Multiple Experts (LFME)
We refer to these models as 'Experts', and the proposed LFME framework aggregates the knowledge from multiple 'Experts' to learn a unified student model.
We conduct extensive experiments and demonstrate that our method is able to achieve superior performances compared to state-of-the-art methods.
arXiv Detail & Related papers (2020-01-06T12:57:36Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.