Efficient Sampling of Stochastic Differential Equations with Positive
Semi-Definite Models
- URL: http://arxiv.org/abs/2303.17109v2
- Date: Wed, 24 May 2023 04:27:04 GMT
- Title: Efficient Sampling of Stochastic Differential Equations with Positive
Semi-Definite Models
- Authors: Anant Raj, Umut \c{S}im\c{s}ekli and Alessandro Rudi
- Abstract summary: This paper deals with the problem of efficient sampling from a differential equation, given the drift function and the diffusion matrix.
It is possible to obtain independent and identically distributed (i.i.d.) samples at precision $varepsilon$ with a cost that is $m2 d log (1/varepsilon)$
Our results suggest that as the true solution gets smoother, we can circumvent the curse of dimensionality without requiring any sort of convexity.
- Score: 91.22420505636006
- License: http://creativecommons.org/publicdomain/zero/1.0/
- Abstract: This paper deals with the problem of efficient sampling from a stochastic
differential equation, given the drift function and the diffusion matrix. The
proposed approach leverages a recent model for probabilities \cite{rudi2021psd}
(the positive semi-definite -- PSD model) from which it is possible to obtain
independent and identically distributed (i.i.d.) samples at precision
$\varepsilon$ with a cost that is $m^2 d \log(1/\varepsilon)$ where $m$ is the
dimension of the model, $d$ the dimension of the space. The proposed approach
consists in: first, computing the PSD model that satisfies the Fokker-Planck
equation (or its fractional variant) associated with the SDE, up to error
$\varepsilon$, and then sampling from the resulting PSD model. Assuming some
regularity of the Fokker-Planck solution (i.e. $\beta$-times differentiability
plus some geometric condition on its zeros) We obtain an algorithm that: (a) in
the preparatory phase obtains a PSD model with L2 distance $\varepsilon$ from
the solution of the equation, with a model of dimension $m =
\varepsilon^{-(d+1)/(\beta-2s)} (\log(1/\varepsilon))^{d+1}$ where $1/2\leq
s\leq1$ is the fractional power to the Laplacian, and total computational
complexity of $O(m^{3.5} \log(1/\varepsilon))$ and then (b) for Fokker-Planck
equation, it is able to produce i.i.d.\ samples with error $\varepsilon$ in
Wasserstein-1 distance, with a cost that is $O(d \varepsilon^{-2(d+1)/\beta-2}
\log(1/\varepsilon)^{2d+3})$ per sample. This means that, if the probability
associated with the SDE is somewhat regular, i.e. $\beta \geq 4d+2$, then the
algorithm requires $O(\varepsilon^{-0.88} \log(1/\varepsilon)^{4.5d})$ in the
preparatory phase, and $O(\varepsilon^{-1/2}\log(1/\varepsilon)^{2d+2})$ for
each sample. Our results suggest that as the true solution gets smoother, we
can circumvent the curse of dimensionality without requiring any sort of
convexity.
Related papers
- Dimension-free Private Mean Estimation for Anisotropic Distributions [55.86374912608193]
Previous private estimators on distributions over $mathRd suffer from a curse of dimensionality.
We present an algorithm whose sample complexity has improved dependence on dimension.
arXiv Detail & Related papers (2024-11-01T17:59:53Z) - Iterative thresholding for non-linear learning in the strong $\varepsilon$-contamination model [3.309767076331365]
We derive approximation bounds for learning single neuron models using thresholded descent.
We also study the linear regression problem, where $sigma(mathbfx) = mathbfx$.
arXiv Detail & Related papers (2024-09-05T16:59:56Z) - ReSQueing Parallel and Private Stochastic Convex Optimization [59.53297063174519]
We introduce a new tool for BFG convex optimization (SCO): a Reweighted Query (ReSQue) estimator for the gradient of a function convolved with a (Gaussian) probability density.
We develop algorithms achieving state-of-the-art complexities for SCO in parallel and private settings.
arXiv Detail & Related papers (2023-01-01T18:51:29Z) - Near Sample-Optimal Reduction-based Policy Learning for Average Reward
MDP [58.13930707612128]
This work considers the sample complexity of obtaining an $varepsilon$-optimal policy in an average reward Markov Decision Process (AMDP)
We prove an upper bound of $widetilde O(H varepsilon-3 ln frac1delta)$ samples per state-action pair, where $H := sp(h*)$ is the span of bias of any optimal policy, $varepsilon$ is the accuracy and $delta$ is the failure probability.
arXiv Detail & Related papers (2022-12-01T15:57:58Z) - Sample-Efficient Reinforcement Learning for Linearly-Parameterized MDPs
with a Generative Model [3.749193647980305]
This paper considers a Markov decision process (MDP) that admits a set of state-action features.
We show that a model-based approach (resp.$$Q-learning) provably learns an $varepsilon$-optimal policy with high probability.
arXiv Detail & Related papers (2021-05-28T17:49:39Z) - Learning elliptic partial differential equations with randomized linear
algebra [2.538209532048867]
We show that one can construct an approximant to $G$ that converges almost surely.
The quantity $0Gamma_epsilonleq 1$ characterizes the quality of the training dataset.
arXiv Detail & Related papers (2021-01-31T16:57:59Z) - SDP Achieves Exact Minimax Optimality in Phase Synchronization [19.909352968029584]
We study the phase synchronization problem with noisy measurements $Y=z*z*+sigma WinmathbbCntimes ntimes n.
We prove that SDP achieves error bound $ (1+o)fracnp22np$ under a squared $ell$ loss.
arXiv Detail & Related papers (2021-01-07T03:14:05Z) - Denoising modulo samples: k-NN regression and tightness of SDP
relaxation [5.025654873456756]
We derive a two-stage algorithm that recovers estimates of the samples $f(x_i)$ with a uniform error rate $O(fraclog nn)frac1d+2)$ holding with high probability.
The estimates of the samples $f(x_i)$ can be subsequently utilized to construct an estimate of the function $f$.
arXiv Detail & Related papers (2020-09-10T13:32:46Z) - Optimal Robust Linear Regression in Nearly Linear Time [97.11565882347772]
We study the problem of high-dimensional robust linear regression where a learner is given access to $n$ samples from the generative model $Y = langle X,w* rangle + epsilon$
We propose estimators for this problem under two settings: (i) $X$ is L4-L2 hypercontractive, $mathbbE [XXtop]$ has bounded condition number and $epsilon$ has bounded variance and (ii) $X$ is sub-Gaussian with identity second moment and $epsilon$ is
arXiv Detail & Related papers (2020-07-16T06:44:44Z) - Model-Free Reinforcement Learning: from Clipped Pseudo-Regret to Sample
Complexity [59.34067736545355]
Given an MDP with $S$ states, $A$ actions, the discount factor $gamma in (0,1)$, and an approximation threshold $epsilon > 0$, we provide a model-free algorithm to learn an $epsilon$-optimal policy.
For small enough $epsilon$, we show an improved algorithm with sample complexity.
arXiv Detail & Related papers (2020-06-06T13:34:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.