Physics-Informed Machine Learning of Dynamical Systems for Efficient
Bayesian Inference
- URL: http://arxiv.org/abs/2209.09349v1
- Date: Mon, 19 Sep 2022 21:17:23 GMT
- Title: Physics-Informed Machine Learning of Dynamical Systems for Efficient
Bayesian Inference
- Authors: Somayajulu L. N. Dhulipala and Yifeng Che and Michael D. Shields
- Abstract summary: No-u-turn sampler (NUTS) is a widely adopted method for performing Bayesian inference.
Hamiltonian neural networks (HNNs) are a noteworthy architecture.
We propose the use of HNNs for performing Bayesian inference efficiently without requiring numerous posterior gradients.
- Score: 0.0
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Although the no-u-turn sampler (NUTS) is a widely adopted method for
performing Bayesian inference, it requires numerous posterior gradients which
can be expensive to compute in practice. Recently, there has been a significant
interest in physics-based machine learning of dynamical (or Hamiltonian)
systems and Hamiltonian neural networks (HNNs) is a noteworthy architecture.
But these types of architectures have not been applied to solve Bayesian
inference problems efficiently. We propose the use of HNNs for performing
Bayesian inference efficiently without requiring numerous posterior gradients.
We introduce latent variable outputs to HNNs (L-HNNs) for improved expressivity
and reduced integration errors. We integrate L-HNNs in NUTS and further propose
an online error monitoring scheme to prevent sampling degeneracy in regions
where L-HNNs may have little training data. We demonstrate L-HNNs in NUTS with
online error monitoring considering several complex high-dimensional posterior
densities and compare its performance to NUTS.
Related papers
- Implicit Stochastic Gradient Descent for Training Physics-informed
Neural Networks [51.92362217307946]
Physics-informed neural networks (PINNs) have effectively been demonstrated in solving forward and inverse differential equation problems.
PINNs are trapped in training failures when the target functions to be approximated exhibit high-frequency or multi-scale features.
In this paper, we propose to employ implicit gradient descent (ISGD) method to train PINNs for improving the stability of training process.
arXiv Detail & Related papers (2023-03-03T08:17:47Z) - Recurrent Bilinear Optimization for Binary Neural Networks [58.972212365275595]
BNNs neglect the intrinsic bilinear relationship of real-valued weights and scale factors.
Our work is the first attempt to optimize BNNs from the bilinear perspective.
We obtain robust RBONNs, which show impressive performance over state-of-the-art BNNs on various models and datasets.
arXiv Detail & Related papers (2022-09-04T06:45:33Z) - Bayesian Inference with Latent Hamiltonian Neural Networks [0.0]
Hamiltonian neural networks (HNNs) with Hamiltonian Monte Carlo (HMC) and No-U-Turn Sampler (NUTS)
HNNs do not require numerical gradients of the target density during sampling.
L-HNNs in NUTS with online error monitoring required 1--2 orders of magnitude fewer numerical gradients of the target density.
arXiv Detail & Related papers (2022-08-12T05:10:18Z) - Training High-Performance Low-Latency Spiking Neural Networks by
Differentiation on Spike Representation [70.75043144299168]
Spiking Neural Network (SNN) is a promising energy-efficient AI model when implemented on neuromorphic hardware.
It is a challenge to efficiently train SNNs due to their non-differentiability.
We propose the Differentiation on Spike Representation (DSR) method, which could achieve high performance.
arXiv Detail & Related papers (2022-05-01T12:44:49Z) - Learning Trajectories of Hamiltonian Systems with Neural Networks [81.38804205212425]
We propose to enhance Hamiltonian neural networks with an estimation of a continuous-time trajectory of the modeled system.
We demonstrate that the proposed integration scheme works well for HNNs, especially with low sampling rates, noisy and irregular observations.
arXiv Detail & Related papers (2022-04-11T13:25:45Z) - Comparative Analysis of Interval Reachability for Robust Implicit and
Feedforward Neural Networks [64.23331120621118]
We use interval reachability analysis to obtain robustness guarantees for implicit neural networks (INNs)
INNs are a class of implicit learning models that use implicit equations as layers.
We show that our approach performs at least as well as, and generally better than, applying state-of-the-art interval bound propagation methods to INNs.
arXiv Detail & Related papers (2022-04-01T03:31:27Z) - Multilevel Bayesian Deep Neural Networks [0.5892638927736115]
We consider inference associated with deep neural networks (DNNs) and in particular, trace-class neural network (TNN) priors.
TNN priors are defined on functions with infinitely many hidden units, and have strongly convergent approximations with finitely many hidden units.
In this paper, we leverage the strong convergence of TNN in order to apply Multilevel Monte Carlo (MLMC) to these models.
arXiv Detail & Related papers (2022-03-24T09:49:27Z) - Hamiltonian Deep Neural Networks Guaranteeing Non-vanishing Gradients by
Design [2.752441514346229]
Vanishing and exploding gradients during weight optimization through backpropagation can be difficult to train.
We propose a general class of Hamiltonian DNNs (H-DNNs) that stem from the discretization of continuous-time Hamiltonian systems.
Our main result is that a broad set of H-DNNs ensures non-vanishing gradients by design for an arbitrary network depth.
The good performance of H-DNNs is demonstrated on benchmark classification problems, including image classification with the MNIST dataset.
arXiv Detail & Related papers (2021-05-27T14:52:22Z) - A unified framework for Hamiltonian deep neural networks [3.0934684265555052]
Training deep neural networks (DNNs) can be difficult due to vanishing/exploding gradients during weight optimization.
We propose a class of DNNs stemming from the time discretization of Hamiltonian systems.
The proposed Hamiltonian framework, besides encompassing existing networks inspired by marginally stable ODEs, allows one to derive new and more expressive architectures.
arXiv Detail & Related papers (2021-04-27T13:20:24Z) - Online Limited Memory Neural-Linear Bandits with Likelihood Matching [53.18698496031658]
We study neural-linear bandits for solving problems where both exploration and representation learning play an important role.
We propose a likelihood matching algorithm that is resilient to catastrophic forgetting and is completely online.
arXiv Detail & Related papers (2021-02-07T14:19:07Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.