Can deep neural networks learn process model structure? An assessment
framework and analysis
- URL: http://arxiv.org/abs/2202.11985v1
- Date: Thu, 24 Feb 2022 09:44:13 GMT
- Title: Can deep neural networks learn process model structure? An assessment
framework and analysis
- Authors: Jari Peeperkorn and Seppe vanden Broucke and Jochen De Weerdt
- Abstract summary: We propose an evaluation scheme complemented with new fitness, precision, and generalisation metrics.
We apply this framework to several process models with simple control-flow behaviour.
Our results show that, even for such simplistic models, careful tuning of overfitting countermeasures is required.
- Score: 0.2580765958706854
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Predictive process monitoring concerns itself with the prediction of ongoing
cases in (business) processes. Prediction tasks typically focus on remaining
time, outcome, next event or full case suffix prediction. Various methods using
machine and deep learning havebeen proposed for these tasks in recent years.
Especially recurrent neural networks (RNNs) such as long short-term memory nets
(LSTMs) have gained in popularity. However, no research focuses on whether such
neural network-based models can truly learn the structure of underlying process
models. For instance, can such neural networks effectively learn parallel
behaviour or loops? Therefore, in this work, we propose an evaluation scheme
complemented with new fitness, precision, and generalisation metrics,
specifically tailored towards measuring the capacity of deep learning models to
learn process model structure. We apply this framework to several process
models with simple control-flow behaviour, on the task of next-event
prediction. Our results show that, even for such simplistic models, careful
tuning of overfitting countermeasures is required to allow these models to
learn process model structure.
Related papers
- Transferable Post-training via Inverse Value Learning [83.75002867411263]
We propose modeling changes at the logits level during post-training using a separate neural network (i.e., the value network)
After training this network on a small base model using demonstrations, this network can be seamlessly integrated with other pre-trained models during inference.
We demonstrate that the resulting value network has broad transferability across pre-trained models of different parameter sizes.
arXiv Detail & Related papers (2024-10-28T13:48:43Z) - Towards Scalable and Versatile Weight Space Learning [51.78426981947659]
This paper introduces the SANE approach to weight-space learning.
Our method extends the idea of hyper-representations towards sequential processing of subsets of neural network weights.
arXiv Detail & Related papers (2024-06-14T13:12:07Z) - A Dynamical Model of Neural Scaling Laws [79.59705237659547]
We analyze a random feature model trained with gradient descent as a solvable model of network training and generalization.
Our theory shows how the gap between training and test loss can gradually build up over time due to repeated reuse of data.
arXiv Detail & Related papers (2024-02-02T01:41:38Z) - Understanding Activation Patterns in Artificial Neural Networks by
Exploring Stochastic Processes [0.0]
We propose utilizing the framework of processes, which has been underutilized thus far.
We focus solely on activation frequency, leveraging neuroscience techniques used for real neuron spike trains.
We derive parameters describing activation patterns in each network, revealing consistent differences across architectures and training sets.
arXiv Detail & Related papers (2023-08-01T22:12:30Z) - How neural networks learn to classify chaotic time series [77.34726150561087]
We study the inner workings of neural networks trained to classify regular-versus-chaotic time series.
We find that the relation between input periodicity and activation periodicity is key for the performance of LKCNN models.
arXiv Detail & Related papers (2023-06-04T08:53:27Z) - Can recurrent neural networks learn process model structure? [0.2580765958706854]
We introduce an evaluation framework that combines variant-based resampling and custom metrics for fitness, precision and generalization.
We confirm that LSTMs can struggle to learn process model structure, even with simplistic process data.
We also found that decreasing the amount of information seen by the LSTM during training, causes a sharp drop in generalization and precision scores.
arXiv Detail & Related papers (2022-12-13T08:40:01Z) - Masking Neural Networks Using Reachability Graphs to Predict Process
Events [0.0]
Decay Replay Mining is a deep learning method that utilizes process model notations to predict the next event.
This paper proposes an approach to further interlock the process model of Replay Mining with its neural network for next event prediction.
arXiv Detail & Related papers (2021-08-01T09:06:55Z) - Gone Fishing: Neural Active Learning with Fisher Embeddings [55.08537975896764]
There is an increasing need for active learning algorithms that are compatible with deep neural networks.
This article introduces BAIT, a practical representation of tractable, and high-performing active learning algorithm for neural networks.
arXiv Detail & Related papers (2021-06-17T17:26:31Z) - Neural Complexity Measures [96.06344259626127]
We propose Neural Complexity (NC), a meta-learning framework for predicting generalization.
Our model learns a scalar complexity measure through interactions with many heterogeneous tasks in a data-driven way.
arXiv Detail & Related papers (2020-08-07T02:12:10Z) - Learning Queuing Networks by Recurrent Neural Networks [0.0]
We propose a machine-learning approach to derive performance models from data.
We exploit a deterministic approximation of their average dynamics in terms of a compact system of ordinary differential equations.
This allows for an interpretable structure of the neural network, which can be trained from system measurements to yield a white-box parameterized model.
arXiv Detail & Related papers (2020-02-25T10:56:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.