Data-Adaptive Probabilistic Likelihood Approximation for Ordinary
Differential Equations
- URL: http://arxiv.org/abs/2306.05566v2
- Date: Thu, 7 Dec 2023 03:27:14 GMT
- Title: Data-Adaptive Probabilistic Likelihood Approximation for Ordinary
Differential Equations
- Authors: Mohan Wu and Martin Lysy
- Abstract summary: We present a novel probabilistic ODE likelihood approximation, DALTON, which can dramatically reduce parameter sensitivity.
Our approximation scales linearly in both ODE variables and time discretization points, and is applicable to ODEs with both partially-unobserved components and non-Gaussian measurement models.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Estimating the parameters of ordinary differential equations (ODEs) is of
fundamental importance in many scientific applications. While ODEs are
typically approximated with deterministic algorithms, new research on
probabilistic solvers indicates that they produce more reliable parameter
estimates by better accounting for numerical errors. However, many ODE systems
are highly sensitive to their parameter values. This produces deep local maxima
in the likelihood function -- a problem which existing probabilistic solvers
have yet to resolve. Here we present a novel probabilistic ODE likelihood
approximation, DALTON, which can dramatically reduce parameter sensitivity by
learning from noisy ODE measurements in a data-adaptive manner. Our
approximation scales linearly in both ODE variables and time discretization
points, and is applicable to ODEs with both partially-unobserved components and
non-Gaussian measurement models. Several examples demonstrate that DALTON
produces more accurate parameter estimates via numerical optimization than
existing probabilistic ODE solvers, and even in some cases than the exact ODE
likelihood itself.
Related papers
- Diffusion Tempering Improves Parameter Estimation with Probabilistic Integrators for Ordinary Differential Equations [34.500484733973536]
Ordinary differential equations (ODEs) are widely used to describe dynamical systems in science, but identifying parameters that explain experimental measurements is challenging.
We propose diffusion tempering, a novel regularization technique for probabilistic numerical methods which improves convergence of gradient-based parameter optimization in ODEs.
We demonstrate that our method is effective for dynamical systems of different complexity and show that it obtains reliable parameter estimates for a Hodgkin-Huxley model with a practically relevant number of parameters.
arXiv Detail & Related papers (2024-02-19T15:36:36Z) - A statistical approach to latent dynamic modeling with differential
equations [0.0]
Ordinary differential equations (ODEs) can provide mechanistic models of temporally local changes of processes.
We propose to use each observation in the course of time as the initial value to obtain multiple local ODE solutions.
We illustrate the proposed approach in an application with spinal muscular atrophy patients and a corresponding simulation study.
arXiv Detail & Related papers (2023-11-27T20:02:55Z) - Probabilistic Unrolling: Scalable, Inverse-Free Maximum Likelihood
Estimation for Latent Gaussian Models [69.22568644711113]
We introduce probabilistic unrolling, a method that combines Monte Carlo sampling with iterative linear solvers to circumvent matrix inversions.
Our theoretical analyses reveal that unrolling and backpropagation through the iterations of the solver can accelerate gradient estimation for maximum likelihood estimation.
In experiments on simulated and real data, we demonstrate that probabilistic unrolling learns latent Gaussian models up to an order of magnitude faster than gradient EM, with minimal losses in model performance.
arXiv Detail & Related papers (2023-06-05T21:08:34Z) - Discovering ordinary differential equations that govern time-series [65.07437364102931]
We propose a transformer-based sequence-to-sequence model that recovers scalar autonomous ordinary differential equations (ODEs) in symbolic form from time-series data of a single observed solution of the ODE.
Our method is efficiently scalable: after one-time pretraining on a large set of ODEs, we can infer the governing laws of a new observed solution in a few forward passes of the model.
arXiv Detail & Related papers (2022-11-05T07:07:58Z) - Identifiability and Asymptotics in Learning Homogeneous Linear ODE Systems from Discrete Observations [114.17826109037048]
Ordinary Differential Equations (ODEs) have recently gained a lot of attention in machine learning.
theoretical aspects, e.g., identifiability and properties of statistical estimation are still obscure.
This paper derives a sufficient condition for the identifiability of homogeneous linear ODE systems from a sequence of equally-spaced error-free observations sampled from a single trajectory.
arXiv Detail & Related papers (2022-10-12T06:46:38Z) - FaDIn: Fast Discretized Inference for Hawkes Processes with General
Parametric Kernels [82.53569355337586]
This work offers an efficient solution to temporal point processes inference using general parametric kernels with finite support.
The method's effectiveness is evaluated by modeling the occurrence of stimuli-induced patterns from brain signals recorded with magnetoencephalography (MEG)
Results show that the proposed approach leads to an improved estimation of pattern latency than the state-of-the-art.
arXiv Detail & Related papers (2022-10-10T12:35:02Z) - Neural Flows: Efficient Alternative to Neural ODEs [8.01886971335823]
We propose an alternative by directly modeling the solution curves - the flow of an ODE - with a neural network.
This immediately eliminates the need for expensive numerical solvers while still maintaining the modeling capability of neural ODEs.
arXiv Detail & Related papers (2021-10-25T15:24:45Z) - A Probabilistic State Space Model for Joint Inference from Differential
Equations and Data [23.449725313605835]
We show a new class of solvers for ordinary differential equations (ODEs) that phrase the solution process directly in terms of Bayesian filtering.
It then becomes possible to perform approximate Bayesian inference on the latent force as well as the ODE solution in a single, linear complexity pass of an extended Kalman filter.
We demonstrate the expressiveness and performance of the algorithm by training a non-parametric SIRD model on data from the COVID-19 outbreak.
arXiv Detail & Related papers (2021-03-18T10:36:09Z) - Meta-Solver for Neural Ordinary Differential Equations [77.8918415523446]
We investigate how the variability in solvers' space can improve neural ODEs performance.
We show that the right choice of solver parameterization can significantly affect neural ODEs models in terms of robustness to adversarial attacks.
arXiv Detail & Related papers (2021-03-15T17:26:34Z) - Stable Implementation of Probabilistic ODE Solvers [27.70274403550477]
Probabilistic solvers for ordinary differential equations (ODEs) provide efficient quantification of numerical uncertainty.
They suffer from numerical instability when run at high order or with small step-sizes.
The present work proposes and examines a solution to this problem.
It involves three components: accurate initialisation, a coordinate change preconditioner that makes numerical stability concerns step-size-independent, and square-root implementation.
arXiv Detail & Related papers (2020-12-18T08:35:36Z) - Optimal oracle inequalities for solving projected fixed-point equations [53.31620399640334]
We study methods that use a collection of random observations to compute approximate solutions by searching over a known low-dimensional subspace of the Hilbert space.
We show how our results precisely characterize the error of a class of temporal difference learning methods for the policy evaluation problem with linear function approximation.
arXiv Detail & Related papers (2020-12-09T20:19:32Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.