Shallow Neural Hawkes: Non-parametric kernel estimation for Hawkes
processes
- URL: http://arxiv.org/abs/2006.02460v1
- Date: Wed, 3 Jun 2020 18:15:38 GMT
- Title: Shallow Neural Hawkes: Non-parametric kernel estimation for Hawkes
processes
- Authors: Sobin Joseph, Lekhapriya Dheeraj Kashyap, Shashi Jain
- Abstract summary: Multi-dimensional Hawkes process (MHP) is a class of self and mutually exciting point processes.
We first find an unbiased estimator for the log-likelihood estimator of the Hawkes process.
We propose a specific single layered neural network for the non-parametric estimation of the underlying kernels.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Multi-dimensional Hawkes process (MHP) is a class of self and mutually
exciting point processes that find wide range of applications -- from
prediction of earthquakes to modelling of order books in high frequency
trading. This paper makes two major contributions, we first find an unbiased
estimator for the log-likelihood estimator of the Hawkes process to enable
efficient use of the stochastic gradient descent method for maximum likelihood
estimation. The second contribution is, we propose a specific single hidden
layered neural network for the non-parametric estimation of the underlying
kernels of the MHP. We evaluate the proposed model on both synthetic and real
datasets, and find the method has comparable or better performance than
existing estimation methods. The use of shallow neural network ensures that we
do not compromise on the interpretability of the Hawkes model, while at the
same time have the flexibility to estimate any non-standard Hawkes excitation
kernel.
Related papers
- Nonparametric estimation of Hawkes processes with RKHSs [1.775610745277615]
This paper addresses nonparametric estimation of nonlinear Hawkes processes, where the interaction functions are assumed to lie in a reproducing kernel space (RKHS)
Motivated by applications in neuroscience, the model allows complex interaction functions, in order to express exciting and inhibiting effects, but also a combination of both.
It shows that our method achieves a better performance compared to related nonparametric estimation techniques and suits neuronal applications.
arXiv Detail & Related papers (2024-11-01T14:26:50Z) - Non-Parametric Estimation of Multi-dimensional Marked Hawkes Processes [0.0]
Marked Hawkes processes feature variable jump size across each event, in contrast to the constant jump size observed in a Hawkes process without marks.
We propose a methodology for estimating the conditional intensity of the marked Hawkes process.
arXiv Detail & Related papers (2024-02-07T10:51:11Z) - Stochastic Marginal Likelihood Gradients using Neural Tangent Kernels [78.6096486885658]
We introduce lower bounds to the linearized Laplace approximation of the marginal likelihood.
These bounds are amenable togradient-based optimization and allow to trade off estimation accuracy against computational complexity.
arXiv Detail & Related papers (2023-06-06T19:02:57Z) - A neural network based model for multi-dimensional nonlinear Hawkes
processes [0.0]
We introduce the Neural Network for Descent Hawkes processes (NNNH), a non-parametric method based on neural networks to fit nonlinear Hawkes processes.
Our results highlight the effectiveness of the NNNH method in accurately capturing the complexities of nonlinear Hawkes processes.
arXiv Detail & Related papers (2023-03-06T12:31:19Z) - Scalable and adaptive variational Bayes methods for Hawkes processes [4.580983642743026]
We propose a novel sparsity-inducing procedure, and derive an adaptive mean-field variational algorithm for the popular sigmoid Hawkes processes.
Our algorithm is parallelisable and therefore computationally efficient in high-dimensional setting.
arXiv Detail & Related papers (2022-12-01T05:35:32Z) - FaDIn: Fast Discretized Inference for Hawkes Processes with General
Parametric Kernels [82.53569355337586]
This work offers an efficient solution to temporal point processes inference using general parametric kernels with finite support.
The method's effectiveness is evaluated by modeling the occurrence of stimuli-induced patterns from brain signals recorded with magnetoencephalography (MEG)
Results show that the proposed approach leads to an improved estimation of pattern latency than the state-of-the-art.
arXiv Detail & Related papers (2022-10-10T12:35:02Z) - NUQ: Nonparametric Uncertainty Quantification for Deterministic Neural
Networks [151.03112356092575]
We show the principled way to measure the uncertainty of predictions for a classifier based on Nadaraya-Watson's nonparametric estimate of the conditional label distribution.
We demonstrate the strong performance of the method in uncertainty estimation tasks on a variety of real-world image datasets.
arXiv Detail & Related papers (2022-02-07T12:30:45Z) - Continuous Wasserstein-2 Barycenter Estimation without Minimax
Optimization [94.18714844247766]
Wasserstein barycenters provide a geometric notion of the weighted average of probability measures based on optimal transport.
We present a scalable algorithm to compute Wasserstein-2 barycenters given sample access to the input measures.
arXiv Detail & Related papers (2021-02-02T21:01:13Z) - Learning Rates as a Function of Batch Size: A Random Matrix Theory
Approach to Neural Network Training [2.9649783577150837]
We study the effect of mini-batching on the loss landscape of deep neural networks using spiked, field-dependent random matrix theory.
We derive analytical expressions for the maximal descent and adaptive training regimens for smooth, non-Newton deep neural networks.
We validate our claims on the VGG/ResNet and ImageNet datasets.
arXiv Detail & Related papers (2020-06-16T11:55:45Z) - Path Sample-Analytic Gradient Estimators for Stochastic Binary Networks [78.76880041670904]
In neural networks with binary activations and or binary weights the training by gradient descent is complicated.
We propose a new method for this estimation problem combining sampling and analytic approximation steps.
We experimentally show higher accuracy in gradient estimation and demonstrate a more stable and better performing training in deep convolutional models.
arXiv Detail & Related papers (2020-06-04T21:51:21Z) - Support recovery and sup-norm convergence rates for sparse pivotal
estimation [79.13844065776928]
In high dimensional sparse regression, pivotal estimators are estimators for which the optimal regularization parameter is independent of the noise level.
We show minimax sup-norm convergence rates for non smoothed and smoothed, single task and multitask square-root Lasso-type estimators.
arXiv Detail & Related papers (2020-01-15T16:11:04Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.