Bayesian Neural Ordinary Differential Equations
- URL: http://arxiv.org/abs/2012.07244v3
- Date: Wed, 3 Mar 2021 16:54:09 GMT
- Title: Bayesian Neural Ordinary Differential Equations
- Authors: Raj Dandekar, Karen Chung, Vaibhav Dixit, Mohamed Tarek, Aslan
Garcia-Valadez, Krishna Vishal Vemula and Chris Rackauckas
- Abstract summary: We demonstrate the successful integration of Neural ODEs with Bayesian inference frameworks.
We achieve a posterior sample accuracy of 98.5% on the test ensemble of 10,000 images.
This gives a scientific machine learning tool for probabilistic estimation of uncertainties.
- Score: 0.9422623204346027
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recently, Neural Ordinary Differential Equations has emerged as a powerful
framework for modeling physical simulations without explicitly defining the
ODEs governing the system, but instead learning them via machine learning.
However, the question: "Can Bayesian learning frameworks be integrated with
Neural ODE's to robustly quantify the uncertainty in the weights of a Neural
ODE?" remains unanswered. In an effort to address this question, we primarily
evaluate the following categories of inference methods: (a) The No-U-Turn MCMC
sampler (NUTS), (b) Stochastic Gradient Hamiltonian Monte Carlo (SGHMC) and (c)
Stochastic Langevin Gradient Descent (SGLD). We demonstrate the successful
integration of Neural ODEs with the above Bayesian inference frameworks on
classical physical systems, as well as on standard machine learning datasets
like MNIST, using GPU acceleration. On the MNIST dataset, we achieve a
posterior sample accuracy of 98.5% on the test ensemble of 10,000 images.
Subsequently, for the first time, we demonstrate the successful integration of
variational inference with normalizing flows and Neural ODEs, leading to a
powerful Bayesian Neural ODE object. Finally, considering a predator-prey model
and an epidemiological system, we demonstrate the probabilistic identification
of model specification in partially-described dynamical systems using universal
ordinary differential equations. Together, this gives a scientific machine
learning tool for probabilistic estimation of epistemic uncertainties.
Related papers
- Learning Chaotic Systems and Long-Term Predictions with Neural Jump ODEs [4.204990010424083]
Pathdependent Neural Jump ODE (PDNJ-ODE) is a model for online prediction of generic processes with irregular (in time) and potentially incomplete (with respect to coordinates) observations.
In this work we enhance the model with two novel ideas, which independently of each other improve the performance of our modelling setup.
The same enhancements can be used to provably enable the PDNJ-ODE to learn long-term predictions for general datasets, where the standard model fails.
arXiv Detail & Related papers (2024-07-26T15:18:29Z) - Foundational Inference Models for Dynamical Systems [5.549794481031468]
We offer a fresh perspective on the classical problem of imputing missing time series data, whose underlying dynamics are assumed to be determined by ODEs.
We propose a novel supervised learning framework for zero-shot time series imputation, through parametric functions satisfying some (hidden) ODEs.
We empirically demonstrate that one and the same (pretrained) recognition model can perform zero-shot imputation across 63 distinct time series with missing values.
arXiv Detail & Related papers (2024-02-12T11:48:54Z) - Capturing dynamical correlations using implicit neural representations [85.66456606776552]
We develop an artificial intelligence framework which combines a neural network trained to mimic simulated data from a model Hamiltonian with automatic differentiation to recover unknown parameters from experimental data.
In doing so, we illustrate the ability to build and train a differentiable model only once, which then can be applied in real-time to multi-dimensional scattering data.
arXiv Detail & Related papers (2023-04-08T07:55:36Z) - Experimental study of Neural ODE training with adaptive solver for
dynamical systems modeling [72.84259710412293]
Some ODE solvers called adaptive can adapt their evaluation strategy depending on the complexity of the problem at hand.
This paper describes a simple set of experiments to show why adaptive solvers cannot be seamlessly leveraged as a black-box for dynamical systems modelling.
arXiv Detail & Related papers (2022-11-13T17:48:04Z) - Identifiability and Asymptotics in Learning Homogeneous Linear ODE Systems from Discrete Observations [114.17826109037048]
Ordinary Differential Equations (ODEs) have recently gained a lot of attention in machine learning.
theoretical aspects, e.g., identifiability and properties of statistical estimation are still obscure.
This paper derives a sufficient condition for the identifiability of homogeneous linear ODE systems from a sequence of equally-spaced error-free observations sampled from a single trajectory.
arXiv Detail & Related papers (2022-10-12T06:46:38Z) - EINNs: Epidemiologically-Informed Neural Networks [75.34199997857341]
We introduce a new class of physics-informed neural networks-EINN-crafted for epidemic forecasting.
We investigate how to leverage both the theoretical flexibility provided by mechanistic models as well as the data-driven expressability afforded by AI models.
arXiv Detail & Related papers (2022-02-21T18:59:03Z) - Mixed Effects Neural ODE: A Variational Approximation for Analyzing the
Dynamics of Panel Data [50.23363975709122]
We propose a probabilistic model called ME-NODE to incorporate (fixed + random) mixed effects for analyzing panel data.
We show that our model can be derived using smooth approximations of SDEs provided by the Wong-Zakai theorem.
We then derive Evidence Based Lower Bounds for ME-NODE, and develop (efficient) training algorithms.
arXiv Detail & Related papers (2022-02-18T22:41:51Z) - Using scientific machine learning for experimental bifurcation analysis
of dynamic systems [2.204918347869259]
This study focuses on training universal differential equation (UDE) models for physical nonlinear dynamical systems with limit cycles.
We consider examples where training data is generated by numerical simulations, whereas we also employ the proposed modelling concept to physical experiments.
We use both neural networks and Gaussian processes as universal approximators alongside the mechanistic models to give a critical assessment of the accuracy and robustness of the UDE modelling approach.
arXiv Detail & Related papers (2021-10-22T15:43:03Z) - Neural Stochastic Partial Differential Equations [1.2183405753834562]
We introduce the Neural SPDE model providing an extension to two important classes of physics-inspired neural architectures.
On the one hand, it extends all the popular neural -- ordinary, controlled, rough -- differential equation models in that it is capable of processing incoming information.
On the other hand, it extends Neural Operators -- recent generalizations of neural networks modelling mappings between functional spaces -- in that it can be used to learn complex SPDE solution operators.
arXiv Detail & Related papers (2021-10-19T20:35:37Z) - Cubature Kalman Filter Based Training of Hybrid Differential Equation
Recurrent Neural Network Physiological Dynamic Models [13.637931956861758]
We show how we can approximate missing ordinary differential equations with known ODEs using a neural network approximation.
Results indicate that this RBSE approach to training the NN parameters yields better outcomes (measurement/state estimation accuracy) than training the neural network with backpropagation.
arXiv Detail & Related papers (2021-10-12T15:38:13Z) - Stochasticity in Neural ODEs: An Empirical Study [68.8204255655161]
Regularization of neural networks (e.g. dropout) is a widespread technique in deep learning that allows for better generalization.
We show that data augmentation during the training improves the performance of both deterministic and versions of the same model.
However, the improvements obtained by the data augmentation completely eliminate the empirical regularization gains, making the performance of neural ODE and neural SDE negligible.
arXiv Detail & Related papers (2020-02-22T22:12:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.