XNAP: Making LSTM-based Next Activity Predictions Explainable by Using
LRP
- URL: http://arxiv.org/abs/2008.07993v3
- Date: Wed, 23 Dec 2020 19:03:05 GMT
- Title: XNAP: Making LSTM-based Next Activity Predictions Explainable by Using
LRP
- Authors: Sven Weinzierl and Sandra Zilker and Jens Brunk and Kate Revoredo and
Martin Matzner and J\"org Becker
- Abstract summary: Predictive business process monitoring (PBPM) is a class of techniques designed to predict behaviour, such as next activities, in running traces.
With the use of deep neural networks (DNNs), the techniques predictive quality could be improved for tasks like the next activity prediction.
In this paper, we propose XNAP, the first explainable, DNN-based PBPM technique for the next activity prediction.
- Score: 0.415623340386296
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Predictive business process monitoring (PBPM) is a class of techniques
designed to predict behaviour, such as next activities, in running traces. PBPM
techniques aim to improve process performance by providing predictions to
process analysts, supporting them in their decision making. However, the PBPM
techniques` limited predictive quality was considered as the essential obstacle
for establishing such techniques in practice. With the use of deep neural
networks (DNNs), the techniques` predictive quality could be improved for tasks
like the next activity prediction. While DNNs achieve a promising predictive
quality, they still lack comprehensibility due to their hierarchical approach
of learning representations. Nevertheless, process analysts need to comprehend
the cause of a prediction to identify intervention mechanisms that might affect
the decision making to secure process performance. In this paper, we propose
XNAP, the first explainable, DNN-based PBPM technique for the next activity
prediction. XNAP integrates a layer-wise relevance propagation method from the
field of explainable artificial intelligence to make predictions of a long
short-term memory DNN explainable by providing relevance values for activities.
We show the benefit of our approach through two real-life event logs.
Related papers
- Learning Logic Specifications for Policy Guidance in POMDPs: an
Inductive Logic Programming Approach [57.788675205519986]
We learn high-quality traces from POMDP executions generated by any solver.
We exploit data- and time-efficient Indu Logic Programming (ILP) to generate interpretable belief-based policy specifications.
We show that learneds expressed in Answer Set Programming (ASP) yield performance superior to neural networks and similar to optimal handcrafted task-specifics within lower computational time.
arXiv Detail & Related papers (2024-02-29T15:36:01Z) - SNAP: Semantic Stories for Next Activity Prediction [4.5723650480442535]
Predicting the next activity in an ongoing process is one of the most common classification tasks in the business process management domain.
Current state-of-the-art AI models for business process prediction do not fully capitalize on available semantic information within process event logs.
We propose a novel SNAP method that leverages language foundation models by constructing semantic contextual stories from the process historical event logs.
arXiv Detail & Related papers (2024-01-28T10:20:15Z) - Knowledge-Driven Modulation of Neural Networks with Attention Mechanism
for Next Activity Prediction [8.552757384215813]
We present a Symbolic[Neuro] system that leverages background knowledge expressed in terms of a procedural process model to offset the under-sampling in the training data.
More specifically, we make predictions using NNs with attention mechanism, an emerging technology in the NN field.
The system has been tested on several real-life logs showing an improvement in the performance of the prediction task.
arXiv Detail & Related papers (2023-12-14T12:02:35Z) - Adversarial Attacks on the Interpretation of Neuron Activation
Maximization [70.5472799454224]
Activation-maximization approaches are used to interpret and analyze trained deep-learning models.
In this work, we consider the concept of an adversary manipulating a model for the purpose of deceiving the interpretation.
arXiv Detail & Related papers (2023-06-12T19:54:33Z) - Interpretable Self-Aware Neural Networks for Robust Trajectory
Prediction [50.79827516897913]
We introduce an interpretable paradigm for trajectory prediction that distributes the uncertainty among semantic concepts.
We validate our approach on real-world autonomous driving data, demonstrating superior performance over state-of-the-art baselines.
arXiv Detail & Related papers (2022-11-16T06:28:20Z) - Scalable PAC-Bayesian Meta-Learning via the PAC-Optimal Hyper-Posterior:
From Theory to Practice [54.03076395748459]
A central question in the meta-learning literature is how to regularize to ensure generalization to unseen tasks.
We present a generalization bound for meta-learning, which was first derived by Rothfuss et al.
We provide a theoretical analysis and empirical case study under which conditions and to what extent these guarantees for meta-learning improve upon PAC-Bayesian per-task learning bounds.
arXiv Detail & Related papers (2022-11-14T08:51:04Z) - MARS: Meta-Learning as Score Matching in the Function Space [79.73213540203389]
We present a novel approach to extracting inductive biases from a set of related datasets.
We use functional Bayesian neural network inference, which views the prior as a process and performs inference in the function space.
Our approach can seamlessly acquire and represent complex prior knowledge by metalearning the score function of the data-generating process.
arXiv Detail & Related papers (2022-10-24T15:14:26Z) - Learning Predictions for Algorithms with Predictions [49.341241064279714]
We introduce a general design approach for algorithms that learn predictors.
We apply techniques from online learning to learn against adversarial instances, tune robustness-consistency trade-offs, and obtain new statistical guarantees.
We demonstrate the effectiveness of our approach at deriving learning algorithms by analyzing methods for bipartite matching, page migration, ski-rental, and job scheduling.
arXiv Detail & Related papers (2022-02-18T17:25:43Z) - Interpreting Process Predictions using a Milestone-Aware Counterfactual
Approach [0.0]
We explore the use of a popular model-agnostic counterfactual algorithm, DiCE, in the context of predictive process analytics.
The analysis reveals that the algorithm is limited when being applied to derive explanations of process predictions.
We propose an approach that supports deriving milestone-aware counterfactuals at different stages of a trace to promote interpretability.
arXiv Detail & Related papers (2021-07-19T09:14:16Z) - Prescriptive Business Process Monitoring for Recommending Next Best
Actions [0.0]
Predictive business process monitoring (PBPM) techniques predict future process behaviour based on historical event log data.
Recent PBPM techniques use state-of-the-art deep neural networks (DNNs) to learn predictive models for producing more accurate predictions.
We present a PrBPM technique that transforms the next most likely activities into the next best actions regarding a given.
arXiv Detail & Related papers (2020-08-19T22:33:54Z) - An empirical comparison of deep-neural-network architectures for next
activity prediction using context-enriched process event logs [0.0]
Researchers have proposed a variety of predictive business process monitoring (PBPM) techniques.
These techniques rely on deep neural networks (DNNs) and consider information about the context, in which the process is running.
We evaluate the predictive quality of three promising DNN architectures, combined with five proven encoding techniques and based on five context-enriched real-life event logs.
arXiv Detail & Related papers (2020-05-03T21:33:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.