Neural Markov Jump Processes
- URL: http://arxiv.org/abs/2305.19744v1
- Date: Wed, 31 May 2023 11:10:29 GMT
- Title: Neural Markov Jump Processes
- Authors: Patrick Seifner and Ramses J. Sanchez
- Abstract summary: We introduce an alternative, variational inference algorithm for Markov jump processes which relies on neural ordinary differential equations.
Our methodology learns neural, continuous-time representations of the observed data, that are used to approximate the initial distribution and time-dependent transition probability rates of the posterior Markov jump process.
We test our approach on synthetic data sampled from ground-truth Markov jump processes, experimental switching ion channel data and molecular dynamics simulations.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Markov jump processes are continuous-time stochastic processes with a wide
range of applications in both natural and social sciences. Despite their
widespread use, inference in these models is highly non-trivial and typically
proceeds via either Monte Carlo or expectation-maximization methods. In this
work we introduce an alternative, variational inference algorithm for Markov
jump processes which relies on neural ordinary differential equations, and is
trainable via back-propagation. Our methodology learns neural, continuous-time
representations of the observed data, that are used to approximate the initial
distribution and time-dependent transition probability rates of the posterior
Markov jump process. The time-independent rates of the prior process are in
contrast trained akin to generative adversarial networks. We test our approach
on synthetic data sampled from ground-truth Markov jump processes, experimental
switching ion channel data and molecular dynamics simulations. Source code to
reproduce our experiments is available online.
Related papers
- Neural MJD: Neural Non-Stationary Merton Jump Diffusion for Time Series Prediction [13.819057582932214]
We introduce Neural MJD, a neural network based non-stationary Merton diffusion (MJD) model.<n>Our model explicitly formulates forecasting as a Poisson equation (SDE) simulation problem.<n>To enable tractable learning, we introduce a likelihood truncation mechanism that caps the number of jumps within small time intervals.
arXiv Detail & Related papers (2025-06-05T01:23:28Z) - High-Dimensional Markov-switching Ordinary Differential Processes [23.17395115394655]
We develop a two-stage algorithm that first recovers the continuous sample path from discrete samples and then estimates the parameters of the processes.
We provide novel theoretical insights into the statistical error and linear convergence guarantee when the processes are $beta$-mixing.
We apply this model to investigate the differences in resting-state brain networks between the ADHD group and normal controls.
arXiv Detail & Related papers (2024-12-30T18:41:28Z) - Logistic-beta processes for dependent random probabilities with beta marginals [58.91121576998588]
We propose a novel process called the logistic-beta process, whose logistic transformation yields a process with common beta marginals.
It can model dependence on both discrete and continuous domains, such as space or time, and has a flexible dependence structure through correlation kernels.
We illustrate the benefits through nonparametric binary regression and conditional density estimation examples, both in simulation studies and in a pregnancy outcome application.
arXiv Detail & Related papers (2024-02-10T21:41:32Z) - Latent State Models of Training Dynamics [51.88132043461152]
We train models with different random seeds and compute a variety of metrics throughout training.
We then fit a hidden Markov model (HMM) over the resulting sequences of metrics.
We use the HMM representation to study phase transitions and identify latent "detour" states that slow down convergence.
arXiv Detail & Related papers (2023-08-18T13:20:08Z) - Real-Time Variational Method for Learning Neural Trajectory and its
Dynamics [7.936841911281107]
We introduce the exponential family variational Kalman filter (eVKF), an online Bayesian method aimed at inferring latent trajectories while simultaneously learning the dynamical system generating them.
We derive a closed-form variational analogue to the predict step of the Kalman filter which leads to a provably tighter bound on the ELBO compared to another online variational method.
We validate our method on synthetic and real-world data, and, notably, show that it achieves competitive performance.
arXiv Detail & Related papers (2023-05-18T19:52:46Z) - Neural Continuous-Time Markov Models [2.28438857884398]
We develop a method to learn a continuous-time Markov chain's transition rate functions from fully observed time series.
We show that our method learns these transition rates with considerably more accuracy than log-linear methods.
arXiv Detail & Related papers (2022-12-11T00:07:41Z) - Score-based Continuous-time Discrete Diffusion Models [102.65769839899315]
We extend diffusion models to discrete variables by introducing a Markov jump process where the reverse process denoises via a continuous-time Markov chain.
We show that an unbiased estimator can be obtained via simple matching the conditional marginal distributions.
We demonstrate the effectiveness of the proposed method on a set of synthetic and real-world music and image benchmarks.
arXiv Detail & Related papers (2022-11-30T05:33:29Z) - Fast Sampling of Diffusion Models via Operator Learning [74.37531458470086]
We use neural operators, an efficient method to solve the probability flow differential equations, to accelerate the sampling process of diffusion models.
Compared to other fast sampling methods that have a sequential nature, we are the first to propose a parallel decoding method.
We show our method achieves state-of-the-art FID of 3.78 for CIFAR-10 and 7.83 for ImageNet-64 in the one-model-evaluation setting.
arXiv Detail & Related papers (2022-11-24T07:30:27Z) - Markov Chain Monte Carlo for Continuous-Time Switching Dynamical Systems [26.744964200606784]
We propose a novel inference algorithm utilizing a Markov Chain Monte Carlo approach.
The presented Gibbs sampler allows to efficiently obtain samples from the exact continuous-time posterior processes.
arXiv Detail & Related papers (2022-05-18T09:03:00Z) - DriPP: Driven Point Processes to Model Stimuli Induced Patterns in M/EEG
Signals [62.997667081978825]
We develop a novel statistical point process model-called driven temporal point processes (DriPP)
We derive a fast and principled expectation-maximization (EM) algorithm to estimate the parameters of this model.
Results on standard MEG datasets demonstrate that our methodology reveals event-related neural responses.
arXiv Detail & Related papers (2021-12-08T13:07:21Z) - Variational Inference for Continuous-Time Switching Dynamical Systems [29.984955043675157]
We present a model based on an Markov jump process modulating a subordinated diffusion process.
We develop a new continuous-time variational inference algorithm.
We extensively evaluate our algorithm under the model assumption and for real-world examples.
arXiv Detail & Related papers (2021-09-29T15:19:51Z) - Nonlinear Independent Component Analysis for Continuous-Time Signals [85.59763606620938]
We study the classical problem of recovering a multidimensional source process from observations of mixtures of this process.
We show that this recovery is possible for many popular models of processes (up to order and monotone scaling of their coordinates) if the mixture is given by a sufficiently differentiable, invertible function.
arXiv Detail & Related papers (2021-02-04T20:28:44Z) - Variational inference formulation for a model-free simulation of a
dynamical system with unknown parameters by a recurrent neural network [8.616180927172548]
We propose a "model-free" simulation of a dynamical system with unknown parameters without prior knowledge.
The deep learning model aims to jointly learn the nonlinear time marching operator and the effects of the unknown parameters from a time series dataset.
It is found that the proposed deep learning model is capable of correctly identifying the dimensions of the random parameters and learning a representation of complex time series data.
arXiv Detail & Related papers (2020-03-02T20:57:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.