Hamiltonian Dynamics with Non-Newtonian Momentum for Rapid Sampling
- URL: http://arxiv.org/abs/2111.02434v1
- Date: Wed, 3 Nov 2021 18:00:07 GMT
- Title: Hamiltonian Dynamics with Non-Newtonian Momentum for Rapid Sampling
- Authors: Greg Ver Steeg and Aram Galstyan
- Abstract summary: Sampling from an unnormalized probability distribution is a fundamental problem in machine learning.
We propose a fundamentally different approach to this problem via a new Hamiltonian dynamics with a non-Newtonian momentum.
In contrast to MCMC approaches like Hamiltonian Monte Carlo, no step is required. Instead, the proposed deterministic dynamics in an extended state space exactly sample the target distribution.
- Score: 38.367354572578314
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Sampling from an unnormalized probability distribution is a fundamental
problem in machine learning with applications including Bayesian modeling,
latent factor inference, and energy-based model training. After decades of
research, variations of MCMC remain the default approach to sampling despite
slow convergence. Auxiliary neural models can learn to speed up MCMC, but the
overhead for training the extra model can be prohibitive. We propose a
fundamentally different approach to this problem via a new Hamiltonian dynamics
with a non-Newtonian momentum. In contrast to MCMC approaches like Hamiltonian
Monte Carlo, no stochastic step is required. Instead, the proposed
deterministic dynamics in an extended state space exactly sample the target
distribution, specified by an energy function, under an assumption of
ergodicity. Alternatively, the dynamics can be interpreted as a normalizing
flow that samples a specified energy model without training. The proposed
Energy Sampling Hamiltonian (ESH) dynamics have a simple form that can be
solved with existing ODE solvers, but we derive a specialized solver that
exhibits much better performance. ESH dynamics converge faster than their MCMC
competitors enabling faster, more stable training of neural network energy
models.
Related papers
- Latent Space Energy-based Neural ODEs [73.01344439786524]
This paper introduces a novel family of deep dynamical models designed to represent continuous-time sequence data.
We train the model using maximum likelihood estimation with Markov chain Monte Carlo.
Experiments on oscillating systems, videos and real-world state sequences (MuJoCo) illustrate that ODEs with the learnable energy-based prior outperform existing counterparts.
arXiv Detail & Related papers (2024-09-05T18:14:22Z) - Repelling-Attracting Hamiltonian Monte Carlo [0.8158530638728501]
We propose a variant of Hamiltonian Monte Carlo, called the Repelling-Attracting Hamiltonian Monte Carlo (RAHMC)
RAHMC involves two stages: a mode-repelling stage to encourage the sampler to move away from regions of high probability density; and, a mode-attracting stage, which facilitates the sampler to find and settle near alternative modes.
arXiv Detail & Related papers (2024-03-07T15:54:55Z) - Learning Energy-Based Prior Model with Diffusion-Amortized MCMC [89.95629196907082]
Common practice of learning latent space EBMs with non-convergent short-run MCMC for prior and posterior sampling is hindering the model from further progress.
We introduce a simple but effective diffusion-based amortization method for long-run MCMC sampling and develop a novel learning algorithm for the latent space EBM based on it.
arXiv Detail & Related papers (2023-10-05T00:23:34Z) - Capturing dynamical correlations using implicit neural representations [85.66456606776552]
We develop an artificial intelligence framework which combines a neural network trained to mimic simulated data from a model Hamiltonian with automatic differentiation to recover unknown parameters from experimental data.
In doing so, we illustrate the ability to build and train a differentiable model only once, which then can be applied in real-time to multi-dimensional scattering data.
arXiv Detail & Related papers (2023-04-08T07:55:36Z) - Model-based Reinforcement Learning with a Hamiltonian Canonical ODE
Network [31.822404891887984]
We introduce Hamiltonian canonical ordinary differential equations into the learning process, which inspires a novel model of neural ordinary differential auto-encoder (NODA)
NODA can model the physical world by nature and is flexible to impose Hamiltonian mechanics.
It can consequentially empower an RL agent with the robust extrapolation using a small amount of samples.
arXiv Detail & Related papers (2022-11-02T08:03:21Z) - Particle Dynamics for Learning EBMs [83.59335980576637]
Energy-based modeling is a promising approach to unsupervised learning, which yields many downstream applications from a single model.
The main difficulty in learning energy-based models with the "contrastive approaches" is the generation of samples from the current energy function at each iteration.
This paper proposes an alternative approach to getting these samples and avoiding crude MCMC sampling from the current model.
arXiv Detail & Related papers (2021-11-26T23:41:07Z) - EBMs Trained with Maximum Likelihood are Generator Models Trained with a
Self-adverserial Loss [6.445605125467574]
We replace Langevin dynamics with deterministic solutions of the associated gradient descent ODE.
We show that reintroducing the noise in the dynamics does not lead to a qualitative change in the behavior.
We thus show that EBM training is effectively a self-adversarial procedure rather than maximum likelihood estimation.
arXiv Detail & Related papers (2021-02-23T15:34:12Z) - No MCMC for me: Amortized sampling for fast and stable training of
energy-based models [62.1234885852552]
Energy-Based Models (EBMs) present a flexible and appealing way to represent uncertainty.
We present a simple method for training EBMs at scale using an entropy-regularized generator to amortize the MCMC sampling.
Next, we apply our estimator to the recently proposed Joint Energy Model (JEM), where we match the original performance with faster and stable training.
arXiv Detail & Related papers (2020-10-08T19:17:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.