Flexible Triggering Kernels for Hawkes Process Modeling
- URL: http://arxiv.org/abs/2202.01869v1
- Date: Thu, 3 Feb 2022 22:02:22 GMT
- Title: Flexible Triggering Kernels for Hawkes Process Modeling
- Authors: Yamac Alican Isik, Connor Davis, Paidamoyo Chapfuwa, Ricardo Henao
- Abstract summary: Recently proposed encoder-decoder structures for modeling Hawkes processes use transformer-inspired architectures.
We introduce an efficient and general encoding of the historical event sequence by replacing the complex (multilayered) attention structures with triggering kernels.
- Score: 11.90725359131405
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recently proposed encoder-decoder structures for modeling Hawkes processes
use transformer-inspired architectures, which encode the history of events via
embeddings and self-attention mechanisms. These models deliver better
prediction and goodness-of-fit than their RNN-based counterparts. However, they
often require high computational and memory complexity requirements and
sometimes fail to adequately capture the triggering function of the underlying
process. So motivated, we introduce an efficient and general encoding of the
historical event sequence by replacing the complex (multilayered) attention
structures with triggering kernels of the observed data. Noting the similarity
between the triggering kernels of a point process and the attention scores, we
use a triggering kernel to replace the weights used to build history
representations. Our estimate for the triggering function is equipped with a
sigmoid gating mechanism that captures local-in-time triggering effects that
are otherwise challenging with standard decaying-over-time kernels. Further,
taking both event type representations and temporal embeddings as inputs, the
model learns the underlying triggering type-time kernel parameters given pairs
of event types. We present experiments on synthetic and real data sets widely
used by competing models, while further including a COVID-19 dataset to
illustrate a scenario where longitudinal covariates are available. Results show
the proposed model outperforms existing approaches while being more efficient
in terms of computational complexity and yielding interpretable results via
direct application of the newly introduced kernel.
Related papers
- Continuous-time convolutions model of event sequences [46.3471121117337]
Event sequences are non-uniform and sparse, making traditional models unsuitable.
We propose COTIC, a method based on an efficient convolution neural network designed to handle the non-uniform occurrence of events over time.
COTIC outperforms existing models in predicting the next event time and type, achieving an average rank of 1.5 compared to 3.714 for the nearest competitor.
arXiv Detail & Related papers (2023-02-13T10:34:51Z) - FaDIn: Fast Discretized Inference for Hawkes Processes with General
Parametric Kernels [82.53569355337586]
This work offers an efficient solution to temporal point processes inference using general parametric kernels with finite support.
The method's effectiveness is evaluated by modeling the occurrence of stimuli-induced patterns from brain signals recorded with magnetoencephalography (MEG)
Results show that the proposed approach leads to an improved estimation of pattern latency than the state-of-the-art.
arXiv Detail & Related papers (2022-10-10T12:35:02Z) - Temporal Attention Augmented Transformer Hawkes Process [4.624987488467739]
We come up with a new kind of Transformer-based Hawkes process model, Temporal Attention Augmented Transformer Hawkes Process (TAA-THP)
We modify the traditional dot-product attention structure, and introduce the temporal encoding into attention structure.
We conduct numerous experiments on a wide range of synthetic and real-life datasets to validate the performance of our proposed TAA-THP model.
arXiv Detail & Related papers (2021-12-29T09:45:23Z) - Neural Spectral Marked Point Processes [18.507050473968985]
We introduce a novel and general neural network-based non-stationary influence kernel for handling complex discrete events.
We demonstrate the superior performance of our proposed method compared with the state-of-the-art on synthetic and real data.
arXiv Detail & Related papers (2021-06-20T23:00:37Z) - Synergetic Learning of Heterogeneous Temporal Sequences for
Multi-Horizon Probabilistic Forecasting [48.8617204809538]
We propose Variational Synergetic Multi-Horizon Network (VSMHN), a novel deep conditional generative model.
To learn complex correlations across heterogeneous sequences, a tailored encoder is devised to combine the advances in deep point processes models and variational recurrent neural networks.
Our model can be trained effectively using variational inference and generates predictions with Monte-Carlo simulation.
arXiv Detail & Related papers (2021-01-31T11:00:55Z) - Deep Neural Dynamic Bayesian Networks applied to EEG sleep spindles
modeling [0.0]
We propose a generative model for single-channel EEG that incorporates the constraints experts actively enforce during visual scoring.
We derive algorithms for exact, tractable inference as a special case of Generalized Expectation Maximization.
We validate the model on three public datasets and provide support that more complex models are able to surpass state-of-the-art detectors.
arXiv Detail & Related papers (2020-10-16T21:48:29Z) - A Multi-Channel Neural Graphical Event Model with Negative Evidence [76.51278722190607]
Event datasets are sequences of events of various types occurring irregularly over the time-line.
We propose a non-parametric deep neural network approach in order to estimate the underlying intensity functions.
arXiv Detail & Related papers (2020-02-21T23:10:50Z) - Transformer Hawkes Process [79.16290557505211]
We propose a Transformer Hawkes Process (THP) model, which leverages the self-attention mechanism to capture long-term dependencies.
THP outperforms existing models in terms of both likelihood and event prediction accuracy by a notable margin.
We provide a concrete example, where THP achieves improved prediction performance for learning multiple point processes when incorporating their relational information.
arXiv Detail & Related papers (2020-02-21T13:48:13Z) - Convolutional Tensor-Train LSTM for Spatio-temporal Learning [116.24172387469994]
We propose a higher-order LSTM model that can efficiently learn long-term correlations in the video sequence.
This is accomplished through a novel tensor train module that performs prediction by combining convolutional features across time.
Our results achieve state-of-the-art performance-art in a wide range of applications and datasets.
arXiv Detail & Related papers (2020-02-21T05:00:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.