Encoder-Decoder Generative Adversarial Nets for Suffix Generation and
Remaining Time Prediction of Business Process Models
- URL: http://arxiv.org/abs/2007.16030v2
- Date: Mon, 19 Oct 2020 09:49:24 GMT
- Title: Encoder-Decoder Generative Adversarial Nets for Suffix Generation and
Remaining Time Prediction of Business Process Models
- Authors: Farbod Taymouri, Marcello La Rosa
- Abstract summary: This paper proposes an encoder-decoder architecture grounded on Generative Adversarial Networks (GANs)
GANs work well with differentiable data such as images, but a suffix is a sequence of categorical items.
We use the Gumbel-Softmax distribution to get a differentiable continuous approximation.
- Score: 0.03807314298073299
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper proposes an encoder-decoder architecture grounded on Generative
Adversarial Networks (GANs), that generates a sequence of activities and their
timestamps in an end-to-end way. GANs work well with differentiable data such
as images. However, a suffix is a sequence of categorical items. To this end,
we use the Gumbel-Softmax distribution to get a differentiable continuous
approximation. The training works by putting one neural network against the
other in a two-player game (hence the "adversarial" nature), which leads to
generating suffixes close to the ground truth. From the experimental evaluation
it emerges that the approach is superior to the baselines in terms of the
accuracy of the predicted suffixes and corresponding remaining times, despite
using a naive feature encoding and only engineering features based on control
flow and events completion time.
Related papers
- An Introduction to Discrete Variational Autoencoders [11.274766489368059]
Variational Autoencoders (VAEs) are a principled approach to unsupervised learning with neural networks.<n>In this tutorial, we provide a rigorous, yet practical, introduction to VAEs in which the latent space is made up of latent variables that follow a categorical distribution.
arXiv Detail & Related papers (2025-05-15T14:33:31Z) - Causal Self-supervised Pretrained Frontend with Predictive Code for Speech Separation [42.63061599979695]
Speech separation (SS) seeks to disentangle a multi-talker speech mixture into single-talker speech streams.
Causal separation models, which rely only on past and present information, offer a promising solution for real-time streaming.
We introduce a novel that is designed to mitigate the mismatch between training and run-time inference by implicitly incorporating future information into causal models.
arXiv Detail & Related papers (2025-04-03T06:18:30Z) - Continuous Speculative Decoding for Autoregressive Image Generation [33.05392461723613]
Continuous-valued Autoregressive (AR) image generation models have demonstrated notable superiority over their discrete-token counterparts.
speculative decoding has proven effective in accelerating Large Language Models (LLMs)
This work generalizes the speculative decoding algorithm from discrete tokens to continuous space.
arXiv Detail & Related papers (2024-11-18T09:19:15Z) - Fourier Head: Helping Large Language Models Learn Complex Probability Distributions [7.074506869260538]
We introduce a neural network layer, constructed using Fourier series, which we can easily substitute for any linear layer if we want the outputs to have a more continuous structure.
We perform extensive analysis on synthetic datasets, as well as on large-scale decision making and time series forecasting tasks.
All of our results support the effectiveness of our proposed Fourier head in scenarios where the underlying data distribution has a natural continuous structure.
arXiv Detail & Related papers (2024-10-29T17:27:58Z) - Diffusion Forcing: Next-token Prediction Meets Full-Sequence Diffusion [61.03681839276652]
Diffusion Forcing is a new training paradigm where a diffusion model is trained to denoise a set of tokens with independent per-token noise levels.
We apply Diffusion Forcing to sequence generative modeling by training a causal next-token prediction model to generate one or several future tokens.
arXiv Detail & Related papers (2024-07-01T15:43:25Z) - DiffCut: Catalyzing Zero-Shot Semantic Segmentation with Diffusion Features and Recursive Normalized Cut [62.63481844384229]
Foundation models have emerged as powerful tools across various domains including language, vision, and multimodal tasks.
In this paper, we use a diffusion UNet encoder as a foundation vision encoder and introduce DiffCut, an unsupervised zero-shot segmentation method.
Our work highlights the remarkably accurate semantic knowledge embedded within diffusion UNet encoders that could then serve as foundation vision encoders for downstream tasks.
arXiv Detail & Related papers (2024-06-05T01:32:31Z) - A Stable, Fast, and Fully Automatic Learning Algorithm for Predictive
Coding Networks [65.34977803841007]
Predictive coding networks are neuroscience-inspired models with roots in both Bayesian statistics and neuroscience.
We show how by simply changing the temporal scheduling of the update rule for the synaptic weights leads to an algorithm that is much more efficient and stable than the original one.
arXiv Detail & Related papers (2022-11-16T00:11:04Z) - ED2LM: Encoder-Decoder to Language Model for Faster Document Re-ranking
Inference [70.36083572306839]
This paper proposes a new training and inference paradigm for re-ranking.
We finetune a pretrained encoder-decoder model using in the form of document to query generation.
We show that this encoder-decoder architecture can be decomposed into a decoder-only language model during inference.
arXiv Detail & Related papers (2022-04-25T06:26:29Z) - Self-Conditioned Generative Adversarial Networks for Image Editing [61.50205580051405]
Generative Adversarial Networks (GANs) are susceptible to bias, learned from either the unbalanced data, or through mode collapse.
We argue that this bias is responsible not only for fairness concerns, but that it plays a key role in the collapse of latent-traversal editing methods when deviating away from the distribution's core.
arXiv Detail & Related papers (2022-02-08T18:08:24Z) - Alleviate Exposure Bias in Sequence Prediction \\ with Recurrent Neural
Networks [47.52214243454995]
A popular strategy to train recurrent neural networks (RNNs) is to take the ground truth as input at each time step.
We propose a fully differentiable training algorithm for RNNs to better capture long-term dependencies.
arXiv Detail & Related papers (2021-03-22T06:15:22Z) - Recent Developments Combining Ensemble Smoother and Deep Generative
Networks for Facies History Matching [58.720142291102135]
This research project focuses on the use of autoencoders networks to construct a continuous parameterization for facies models.
We benchmark seven different formulations, including VAE, generative adversarial network (GAN), Wasserstein GAN, variational auto-encoding GAN, principal component analysis (PCA) with cycle GAN, PCA with transfer style network and VAE with style loss.
arXiv Detail & Related papers (2020-05-08T21:32:42Z) - A game-theoretic approach for Generative Adversarial Networks [2.995087247817663]
Generative adversarial networks (GANs) are a class of generative models, known for producing accurate samples.
Main bottleneck for their implementation is that the neural networks are very hard to train.
We propose a relaxed forward-backward algorithm for GANs.
We prove that when the pseudogradient mapping of the game is monotone, we have convergence to an exact solution or in a neighbourhood of it.
arXiv Detail & Related papers (2020-03-30T17:14:41Z) - PushNet: Efficient and Adaptive Neural Message Passing [1.9121961872220468]
Message passing neural networks have recently evolved into a state-of-the-art approach to representation learning on graphs.
Existing methods perform synchronous message passing along all edges in multiple subsequent rounds.
We consider a novel asynchronous message passing approach where information is pushed only along the most relevant edges until convergence.
arXiv Detail & Related papers (2020-03-04T18:15:30Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.