Data-driven Efficient Solvers for Langevin Dynamics on Manifold in High
Dimensions
- URL: http://arxiv.org/abs/2005.12787v3
- Date: Tue, 27 Sep 2022 06:52:48 GMT
- Title: Data-driven Efficient Solvers for Langevin Dynamics on Manifold in High
Dimensions
- Authors: Yuan Gao, Jian-Guo Liu, Nan Wu
- Abstract summary: We study the Langevin dynamics of a physical system with manifold structure $mathcalMsubsetmathbbRp$
We leverage the corresponding Fokker-Planck equation on the manifold $mathcalN$ in terms of the reaction coordinates $mathsfy$.
We propose an implementable, unconditionally stable, data-driven finite volume scheme for this Fokker-Planck equation.
- Score: 12.005576001523515
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the Langevin dynamics of a physical system with manifold structure
$\mathcal{M}\subset\mathbb{R}^p$ based on collected sample points
$\{\mathsf{x}_i\}_{i=1}^n \subset \mathcal{M}$ that probe the unknown manifold
$\mathcal{M}$. Through the diffusion map, we first learn the reaction
coordinates $\{\mathsf{y}_i\}_{i=1}^n\subset \mathcal{N}$ corresponding to
$\{\mathsf{x}_i\}_{i=1}^n$, where $\mathcal{N}$ is a manifold diffeomorphic to
$\mathcal{M}$ and isometrically embedded in $\mathbb{R}^\ell$ with $\ell \ll
p$. The induced Langevin dynamics on $\mathcal{N}$ in terms of the reaction
coordinates captures the slow time scale dynamics such as conformational
changes in biochemical reactions. To construct an efficient and stable
approximation for the Langevin dynamics on $\mathcal{N}$, we leverage the
corresponding Fokker-Planck equation on the manifold $\mathcal{N}$ in terms of
the reaction coordinates $\mathsf{y}$. We propose an implementable,
unconditionally stable, data-driven finite volume scheme for this Fokker-Planck
equation, which automatically incorporates the manifold structure of
$\mathcal{N}$. Furthermore, we provide a weighted $L^2$ convergence analysis of
the finite volume scheme to the Fokker-Planck equation on $\mathcal{N}$. The
proposed finite volume scheme leads to a Markov chain on
$\{\mathsf{y}_i\}_{i=1}^n$ with an approximated transition probability and jump
rate between the nearest neighbor points. After an unconditionally stable
explicit time discretization, the data-driven finite volume scheme gives an
approximated Markov process for the Langevin dynamics on $\mathcal{N}$ and the
approximated Markov process enjoys detailed balance, ergodicity, and other good
properties.
Related papers
- Efficient Sampling on Riemannian Manifolds via Langevin MCMC [51.825900634131486]
We study the task efficiently sampling from a Gibbs distribution $d pi* = eh d vol_g$ over aian manifold $M$ via (geometric) Langevin MCMC.
Our results apply to general settings where $pi*$ can be non exponential and $Mh$ can have negative Ricci curvature.
arXiv Detail & Related papers (2024-02-15T22:59:14Z) - Provably learning a multi-head attention layer [55.2904547651831]
Multi-head attention layer is one of the key components of the transformer architecture that sets it apart from traditional feed-forward models.
In this work, we initiate the study of provably learning a multi-head attention layer from random examples.
We prove computational lower bounds showing that in the worst case, exponential dependence on $m$ is unavoidable.
arXiv Detail & Related papers (2024-02-06T15:39:09Z) - A Unified Framework for Uniform Signal Recovery in Nonlinear Generative
Compressed Sensing [68.80803866919123]
Under nonlinear measurements, most prior results are non-uniform, i.e., they hold with high probability for a fixed $mathbfx*$ rather than for all $mathbfx*$ simultaneously.
Our framework accommodates GCS with 1-bit/uniformly quantized observations and single index models as canonical examples.
We also develop a concentration inequality that produces tighter bounds for product processes whose index sets have low metric entropy.
arXiv Detail & Related papers (2023-09-25T17:54:19Z) - Convergence of a Normal Map-based Prox-SGD Method under the KL
Inequality [0.0]
We present a novel map-based algorithm ($mathsfnorMtext-mathsfSGD$) for $symbol$k$ convergence problems.
arXiv Detail & Related papers (2023-05-10T01:12:11Z) - Learning a Single Neuron with Adversarial Label Noise via Gradient
Descent [50.659479930171585]
We study a function of the form $mathbfxmapstosigma(mathbfwcdotmathbfx)$ for monotone activations.
The goal of the learner is to output a hypothesis vector $mathbfw$ that $F(mathbbw)=C, epsilon$ with high probability.
arXiv Detail & Related papers (2022-06-17T17:55:43Z) - Markovian Repeated Interaction Quantum Systems [0.0]
We study a class of dynamical semigroups $(mathbbLn)_ninmathbbN$ that emerge, by a Feynman--Kac type formalism, from a random quantum dynamical system.
As a physical application, we consider the case where the $mathcalL_omega$'s are the reduced dynamical maps describing the repeated interactions of a system with thermal probes.
arXiv Detail & Related papers (2022-02-10T20:52:40Z) - Metric Hypertransformers are Universal Adapted Maps [4.83420384410068]
metric hypertransformers (MHTs) are capable of approxing any adapted map $F:mathscrXmathbbZrightarrow mathscrYmathbbZ$ with approximable complexity.
Our results provide the first (quantimat) universal approximation theorem compatible with any such $mathscrX$ and $mathscrY$.
arXiv Detail & Related papers (2022-01-31T10:03:46Z) - Multimeasurement Generative Models [7.502947376736449]
We map the problem of sampling from an unknown distribution with density $p_X$ in $mathbbRd$ to the problem of learning and sampling $p_mathbfY$ in $mathbbRMd$ obtained by convolving $p_X$ with a fixed factorial kernel.
arXiv Detail & Related papers (2021-12-18T02:11:36Z) - Approximating the Riemannian Metric from Point Clouds via Manifold
Moving Least Squares [2.2774471443318753]
We present a naive algorithm that yields approximate geodesic distances with a rate of convergence $ O(h) $ provable approximations.
We show the potential and the robustness to noise of the proposed method on some numerical simulations.
arXiv Detail & Related papers (2020-07-20T04:42:17Z) - Linear Time Sinkhorn Divergences using Positive Features [51.50788603386766]
Solving optimal transport with an entropic regularization requires computing a $ntimes n$ kernel matrix that is repeatedly applied to a vector.
We propose to use instead ground costs of the form $c(x,y)=-logdotpvarphi(x)varphi(y)$ where $varphi$ is a map from the ground space onto the positive orthant $RRr_+$, with $rll n$.
arXiv Detail & Related papers (2020-06-12T10:21:40Z) - Agnostic Learning of a Single Neuron with Gradient Descent [92.7662890047311]
We consider the problem of learning the best-fitting single neuron as measured by the expected square loss.
For the ReLU activation, our population risk guarantee is $O(mathsfOPT1/2)+epsilon$.
For the ReLU activation, our population risk guarantee is $O(mathsfOPT1/2)+epsilon$.
arXiv Detail & Related papers (2020-05-29T07:20:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.