Probability flow solution of the Fokker-Planck equation
- URL: http://arxiv.org/abs/2206.04642v1
- Date: Thu, 9 Jun 2022 17:37:09 GMT
- Title: Probability flow solution of the Fokker-Planck equation
- Authors: Nicholas M. Boffi and Eric Vanden-Eijnden
- Abstract summary: We introduce an alternative scheme based on integrating an ordinary differential equation that describes the flow of probability.
Unlike the dynamics, this equation deterministically pushes samples from the initial density onto samples from the solution at any later time.
Our approach is based on recent advances in score-based diffusion for generative modeling.
- Score: 10.484851004093919
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The method of choice for integrating the time-dependent Fokker-Planck
equation in high-dimension is to generate samples from the solution via
integration of the associated stochastic differential equation. Here, we
introduce an alternative scheme based on integrating an ordinary differential
equation that describes the flow of probability. Unlike the stochastic
dynamics, this equation deterministically pushes samples from the initial
density onto samples from the solution at any later time. The method has the
advantage of giving direct access to quantities that are challenging to
estimate only given samples from the solution, such as the probability current,
the density itself, and its entropy. The probability flow equation depends on
the gradient of the logarithm of the solution (its "score"), and so is a-priori
unknown. To resolve this dependence, we model the score with a deep neural
network that is learned on-the-fly by propagating a set of particles according
to the instantaneous probability current. Our approach is based on recent
advances in score-based diffusion for generative modeling, with the important
difference that the training procedure is self-contained and does not require
samples from the target density to be available beforehand. To demonstrate the
validity of the approach, we consider several examples from the physics of
interacting particle systems; we find that the method scales well to
high-dimensional systems, and accurately matches available analytical solutions
and moments computed via Monte-Carlo.
Related papers
- Harmonic Path Integral Diffusion [0.4527270266697462]
We present a novel approach for sampling from a continuous multivariate probability distribution.
Our method constructs a time-dependent bridge from a delta function centered at the origin of the state space.
arXiv Detail & Related papers (2024-09-23T16:20:21Z) - Total Uncertainty Quantification in Inverse PDE Solutions Obtained with Reduced-Order Deep Learning Surrogate Models [50.90868087591973]
We propose an approximate Bayesian method for quantifying the total uncertainty in inverse PDE solutions obtained with machine learning surrogate models.
We test the proposed framework by comparing it with the iterative ensemble smoother and deep ensembling methods for a non-linear diffusion equation.
arXiv Detail & Related papers (2024-08-20T19:06:02Z) - Dynamical Measure Transport and Neural PDE Solvers for Sampling [77.38204731939273]
We tackle the task of sampling from a probability density as transporting a tractable density function to the target.
We employ physics-informed neural networks (PINNs) to approximate the respective partial differential equations (PDEs) solutions.
PINNs allow for simulation- and discretization-free optimization and can be trained very efficiently.
arXiv Detail & Related papers (2024-07-10T17:39:50Z) - Weak Generative Sampler to Efficiently Sample Invariant Distribution of Stochastic Differential Equation [8.67581853745823]
Current deep learning-based method solves the stationary Fokker--Planck equation to determine the invariant probability density function in form of deep neural networks.
We introduce a framework that employs a weak generative sampler (WGS) to directly generate independent and identically distributed (iid) samples.
Our proposed loss function is based on the weak form of the Fokker--Planck equation, integrating normalizing flows to characterize the invariant distribution.
arXiv Detail & Related papers (2024-05-29T16:41:42Z) - Sobolev Space Regularised Pre Density Models [51.558848491038916]
We propose a new approach to non-parametric density estimation that is based on regularizing a Sobolev norm of the density.
This method is statistically consistent, and makes the inductive validation model clear and consistent.
arXiv Detail & Related papers (2023-07-25T18:47:53Z) - MESSY Estimation: Maximum-Entropy based Stochastic and Symbolic densitY
Estimation [4.014524824655106]
MESSY estimation is a Maximum-Entropy based Gradient and Symbolic densitY estimation method.
We construct a gradient-based drift-diffusion process that connects samples of the unknown distribution function to a guess symbolic expression.
We find that the addition of a symbolic search for basis functions improves the accuracy of the estimation at a reasonable additional computational cost.
arXiv Detail & Related papers (2023-06-07T03:28:47Z) - Stochastic Interpolants: A Unifying Framework for Flows and Diffusions [16.95541777254722]
A class of generative models that unifies flow-based and diffusion-based methods is introduced.
These models extend the framework proposed in Albergo & VandenEijnden (2023), enabling the use of a broad class of continuous-time processes called stochastic interpolants'
These interpolants are built by combining data from the two prescribed densities with an additional latent variable that shapes the bridge in a flexible way.
arXiv Detail & Related papers (2023-03-15T17:43:42Z) - Monte Carlo Neural PDE Solver for Learning PDEs via Probabilistic Representation [59.45669299295436]
We propose a Monte Carlo PDE solver for training unsupervised neural solvers.
We use the PDEs' probabilistic representation, which regards macroscopic phenomena as ensembles of random particles.
Our experiments on convection-diffusion, Allen-Cahn, and Navier-Stokes equations demonstrate significant improvements in accuracy and efficiency.
arXiv Detail & Related papers (2023-02-10T08:05:19Z) - Score-based Continuous-time Discrete Diffusion Models [102.65769839899315]
We extend diffusion models to discrete variables by introducing a Markov jump process where the reverse process denoises via a continuous-time Markov chain.
We show that an unbiased estimator can be obtained via simple matching the conditional marginal distributions.
We demonstrate the effectiveness of the proposed method on a set of synthetic and real-world music and image benchmarks.
arXiv Detail & Related papers (2022-11-30T05:33:29Z) - A blob method method for inhomogeneous diffusion with applications to
multi-agent control and sampling [0.6562256987706128]
We develop a deterministic particle method for the weighted porous medium equation (WPME) and prove its convergence on bounded time intervals.
Our method has natural applications to multi-agent coverage algorithms and sampling probability measures.
arXiv Detail & Related papers (2022-02-25T19:49:05Z) - Large-Scale Wasserstein Gradient Flows [84.73670288608025]
We introduce a scalable scheme to approximate Wasserstein gradient flows.
Our approach relies on input neural networks (ICNNs) to discretize the JKO steps.
As a result, we can sample from the measure at each step of the gradient diffusion and compute its density.
arXiv Detail & Related papers (2021-06-01T19:21:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.