Leveraging Data Augmentation and Siamese Learning for Predictive Process Monitoring
- URL: http://arxiv.org/abs/2507.18293v1
- Date: Thu, 24 Jul 2025 10:57:20 GMT
- Title: Leveraging Data Augmentation and Siamese Learning for Predictive Process Monitoring
- Authors: Sjoerd van Straten, Alessandro Padella, Marwan Hassani,
- Abstract summary: SiamSA-PPM is a self-supervised learning framework that combines Siamese learning with Statistical Augmentation for Predictive Process Monitoring.<n>We show that SiamSA-PPM achieves competitive or superior performance compared to the SOTA in both next activity and final outcome prediction tasks.
- Score: 45.023332704223755
- License: http://creativecommons.org/licenses/by-nc-nd/4.0/
- Abstract: Predictive Process Monitoring (PPM) enables forecasting future events or outcomes of ongoing business process instances based on event logs. However, deep learning PPM approaches are often limited by the low variability and small size of real-world event logs. To address this, we introduce SiamSA-PPM, a novel self-supervised learning framework that combines Siamese learning with Statistical Augmentation for Predictive Process Monitoring. It employs three novel statistically grounded transformation methods that leverage control-flow semantics and frequent behavioral patterns to generate realistic, semantically valid new trace variants. These augmented views are used within a Siamese learning setup to learn generalizable representations of process prefixes without the need for labeled supervision. Extensive experiments on real-life event logs demonstrate that SiamSA-PPM achieves competitive or superior performance compared to the SOTA in both next activity and final outcome prediction tasks. Our results further show that statistical augmentation significantly outperforms random transformations and improves variability in the data, highlighting SiamSA-PPM as a promising direction for training data enrichment in process prediction.
Related papers
- Incremental Uncertainty-aware Performance Monitoring with Active Labeling Intervention [64.12447263206381]
We propose Incremental Uncertainty-aware Performance Monitoring (IUPM), a label-free method that estimates performance changes by modeling gradual shifts using optimal transport.<n>IUPM quantifies the uncertainty in the performance prediction and introduces an active labeling procedure to restore a reliable estimate under a limited labeling budget.<n>Our experiments show that IUPM outperforms existing performance estimation baselines in various gradual shift scenarios.
arXiv Detail & Related papers (2025-05-11T15:35:55Z) - Deciphering Cross-Modal Alignment in Large Vision-Language Models with Modality Integration Rate [118.37653302885607]
We present the Modality Integration Rate (MIR), an effective, robust, and generalized metric to indicate the multi-modal pre-training quality of Large Vision Language Models (LVLMs)
MIR is indicative about training data selection, training strategy schedule, and model architecture design to get better pre-training results.
arXiv Detail & Related papers (2024-10-09T17:59:04Z) - Knowledge-Driven Modulation of Neural Networks with Attention Mechanism
for Next Activity Prediction [8.552757384215813]
We present a Symbolic[Neuro] system that leverages background knowledge expressed in terms of a procedural process model to offset the under-sampling in the training data.
More specifically, we make predictions using NNs with attention mechanism, an emerging technology in the NN field.
The system has been tested on several real-life logs showing an improvement in the performance of the prediction task.
arXiv Detail & Related papers (2023-12-14T12:02:35Z) - TiMix: Text-aware Image Mixing for Effective Vision-Language
Pre-training [42.142924806184425]
Mixed data samples for cross-modal contrastive learning implicitly serve as a regularizer for the contrastive loss.
TiMix exhibits a comparable performance on downstream tasks, even with a reduced amount of training data and shorter training time, when benchmarked against existing methods.
arXiv Detail & Related papers (2023-12-14T12:02:24Z) - Performance-Preserving Event Log Sampling for Predictive Monitoring [0.3425341633647624]
We propose an instance selection procedure that allows sampling training process instances for prediction models.
We show that our instance selection procedure allows for a significant increase of training speed for next activity and remaining time prediction methods.
arXiv Detail & Related papers (2023-01-18T16:07:56Z) - Towards Out-of-Distribution Sequential Event Prediction: A Causal
Treatment [72.50906475214457]
The goal of sequential event prediction is to estimate the next event based on a sequence of historical events.
In practice, the next-event prediction models are trained with sequential data collected at one time.
We propose a framework with hierarchical branching structures for learning context-specific representations.
arXiv Detail & Related papers (2022-10-24T07:54:13Z) - Towards Sequence-Level Training for Visual Tracking [60.95799261482857]
This work introduces a sequence-level training strategy for visual tracking based on reinforcement learning.
Four representative tracking models, SiamRPN++, SiamAttn, TransT, and TrDiMP, consistently improve by incorporating the proposed methods in training.
arXiv Detail & Related papers (2022-08-11T13:15:36Z) - Multi-Augmentation for Efficient Visual Representation Learning for
Self-supervised Pre-training [1.3733988835863333]
We propose Multi-Augmentations for Self-Supervised Learning (MA-SSRL), which fully searched for various augmentation policies to build the entire pipeline.
MA-SSRL successfully learns the invariant feature representation and presents an efficient, effective, and adaptable data augmentation pipeline for self-supervised pre-training.
arXiv Detail & Related papers (2022-05-24T04:18:39Z) - Event Log Sampling for Predictive Monitoring [0.3425341633647624]
This paper proposes an instance selection procedure that allows sampling training process instances for prediction models.
We show that our sampling method allows for a significant increase of training speed for next activity prediction methods while maintaining reliable levels of prediction accuracy.
arXiv Detail & Related papers (2022-04-04T13:36:48Z) - Video Prediction via Example Guidance [156.08546987158616]
In video prediction tasks, one major challenge is to capture the multi-modal nature of future contents and dynamics.
In this work, we propose a simple yet effective framework that can efficiently predict plausible future states.
arXiv Detail & Related papers (2020-07-03T14:57:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.