Goal-Oriented Next Best Activity Recommendation using Reinforcement
Learning
- URL: http://arxiv.org/abs/2205.03219v1
- Date: Fri, 6 May 2022 13:48:14 GMT
- Title: Goal-Oriented Next Best Activity Recommendation using Reinforcement
Learning
- Authors: Prerna Agarwal, Avani Gupta, Renuka Sindhgatta, Sampath Dechu
- Abstract summary: We propose a goal-oriented next best activity recommendation framework.
A deep learning model predicts the next best activity and an estimated value of a goal given the activity.
A reinforcement learning method explores the sequence of activities based on the estimates likely to meet one or more goals.
- Score: 4.128679340077271
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recommending a sequence of activities for an ongoing case requires that the
recommendations conform to the underlying business process and meet the
performance goal of either completion time or process outcome. Existing work on
next activity prediction can predict the future activity but cannot provide
guarantees of the prediction being conformant or meeting the goal. Hence, we
propose a goal-oriented next best activity recommendation. Our proposed
framework uses a deep learning model to predict the next best activity and an
estimated value of a goal given the activity. A reinforcement learning method
explores the sequence of activities based on the estimates likely to meet one
or more goals. We further address a real-world problem of multiple goals by
introducing an additional reward function to balance the outcome of a
recommended activity and satisfy the goal. We demonstrate the effectiveness of
the proposed method on four real-world datasets with different characteristics.
The results show that the recommendations from our proposed approach outperform
in goal satisfaction and conformance compared to the existing state-of-the-art
next best activity recommendation techniques.
Related papers
- Human Action Anticipation: A Survey [86.415721659234]
The literature on behavior prediction spans various tasks, including action anticipation, activity forecasting, intent prediction, goal prediction, and so on.
Our survey aims to tie together this fragmented literature, covering recent technical innovations as well as the development of new large-scale datasets for model training and evaluation.
arXiv Detail & Related papers (2024-10-17T21:37:40Z) - Propose, Assess, Search: Harnessing LLMs for Goal-Oriented Planning in Instructional Videos [48.15438373870542]
VidAssist is an integrated framework designed for zero/few-shot goal-oriented planning in instructional videos.
It employs a breadth-first search algorithm for optimal plan generation.
Experiments demonstrate that VidAssist offers a unified framework for different goal-oriented planning setups.
arXiv Detail & Related papers (2024-09-30T17:57:28Z) - From Recognition to Prediction: Leveraging Sequence Reasoning for Action Anticipation [30.161471749050833]
We propose a novel end-to-end video modeling architecture that utilizes attention mechanisms, named Anticipation via Recognition and Reasoning (ARR)
ARR decomposes the action anticipation task into action recognition and reasoning tasks, and effectively learns the statistical relationship between actions by next action prediction (NAP)
In addition, to address the challenge of relationship modeling that requires extensive training data, we propose an innovative approach for the unsupervised pre-training of the decoder.
arXiv Detail & Related papers (2024-08-05T18:38:29Z) - Code Models are Zero-shot Precondition Reasoners [83.8561159080672]
We use code representations to reason about action preconditions for sequential decision making tasks.
We propose a precondition-aware action sampling strategy that ensures actions predicted by a policy are consistent with preconditions.
arXiv Detail & Related papers (2023-11-16T06:19:27Z) - Action Anticipation with Goal Consistency [19.170733994203367]
We propose to harness high-level intent information to anticipate actions that will take place in the future.
We show the effectiveness of the proposed approach and demonstrate that our method achieves state-of-the-art results on two large-scale datasets.
arXiv Detail & Related papers (2023-06-26T20:04:23Z) - Towards Out-of-Distribution Sequential Event Prediction: A Causal
Treatment [72.50906475214457]
The goal of sequential event prediction is to estimate the next event based on a sequence of historical events.
In practice, the next-event prediction models are trained with sequential data collected at one time.
We propose a framework with hierarchical branching structures for learning context-specific representations.
arXiv Detail & Related papers (2022-10-24T07:54:13Z) - Learning to act: a Reinforcement Learning approach to recommend the best
next activities [4.511664266033014]
This paper investigates an approach that learns, by means of Reinforcement Learning, an optimal policy from the observation of past executions.
The potentiality of the approach has been demonstrated on two scenarios taken from real-life data.
arXiv Detail & Related papers (2022-03-29T09:43:39Z) - Learning to Plan Optimistically: Uncertainty-Guided Deep Exploration via
Latent Model Ensembles [73.15950858151594]
This paper presents Latent Optimistic Value Exploration (LOVE), a strategy that enables deep exploration through optimism in the face of uncertain long-term rewards.
We combine latent world models with value function estimation to predict infinite-horizon returns and recover associated uncertainty via ensembling.
We apply LOVE to visual robot control tasks in continuous action spaces and demonstrate on average more than 20% improved sample efficiency in comparison to state-of-the-art and other exploration objectives.
arXiv Detail & Related papers (2020-10-27T22:06:57Z) - Reward Maximisation through Discrete Active Inference [1.2074552857379273]
We show how and when active inference agents perform actions that are optimal for maximising reward.
We show the conditions under which active inference produces the optimal solution to the Bellman equation.
We append the analysis with a discussion of the broader relationship between active inference and reinforcement learning.
arXiv Detail & Related papers (2020-09-17T07:13:59Z) - Long-Term Anticipation of Activities with Cycle Consistency [90.79357258104417]
We propose a framework for anticipating future activities directly from the features of the observed frames and train it in an end-to-end fashion.
Our framework achieves state-the-art results on two datasets: the Breakfast dataset and 50Salads.
arXiv Detail & Related papers (2020-09-02T15:41:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.