Moving Target Monte Carlo
- URL: http://arxiv.org/abs/2003.04873v1
- Date: Tue, 10 Mar 2020 17:38:36 GMT
- Title: Moving Target Monte Carlo
- Authors: Haoyun Ying, Keheng Mao, Klaus Mosegaard
- Abstract summary: We introduce a new non-Markovian sampling algorithm called Moving Target Monte Carlo (MTMC)
The acceptance rate at $n$-th iteration is constructed using an iteratively updated approximation of the posterior distribution $a_n(mathbfx)$.
The true value of the posterior $p(mathbfx|mathbfd)$ is only calculated if the candidate $mathbfx$ is accepted.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The Markov Chain Monte Carlo (MCMC) methods are popular when considering
sampling from a high-dimensional random variable $\mathbf{x}$ with possibly
unnormalised probability density $p$ and observed data $\mathbf{d}$. However,
MCMC requires evaluating the posterior distribution $p(\mathbf{x}|\mathbf{d})$
of the proposed candidate $\mathbf{x}$ at each iteration when constructing the
acceptance rate. This is costly when such evaluations are intractable. In this
paper, we introduce a new non-Markovian sampling algorithm called Moving Target
Monte Carlo (MTMC). The acceptance rate at $n$-th iteration is constructed
using an iteratively updated approximation of the posterior distribution
$a_n(\mathbf{x})$ instead of $p(\mathbf{x}|\mathbf{d})$. The true value of the
posterior $p(\mathbf{x}|\mathbf{d})$ is only calculated if the candidate
$\mathbf{x}$ is accepted. The approximation $a_n$ utilises these evaluations
and converges to $p$ as $n \rightarrow \infty$. A proof of convergence and
estimation of convergence rate in different situations are given.
Related papers
- Optimal Sketching for Residual Error Estimation for Matrix and Vector Norms [50.15964512954274]
We study the problem of residual error estimation for matrix and vector norms using a linear sketch.
We demonstrate that this gives a substantial advantage empirically, for roughly the same sketch size and accuracy as in previous work.
We also show an $Omega(k2/pn1-2/p)$ lower bound for the sparse recovery problem, which is tight up to a $mathrmpoly(log n)$ factor.
arXiv Detail & Related papers (2024-08-16T02:33:07Z) - Finite-Sample Symmetric Mean Estimation with Fisher Information Rate [15.802475232604667]
Mean of an unknown variance-$sigma2$ distribution can be estimated from $n$ samples with variance $fracsigma2n$ and nearly corresponding subgaussian rate.
When $f$ is known up to translation, this can be improved to $frac1nmathcal I$, where $mathcal I$ is the Fisher information of the distribution.
arXiv Detail & Related papers (2023-06-28T21:31:46Z) - Fast $(1+\varepsilon)$-Approximation Algorithms for Binary Matrix
Factorization [54.29685789885059]
We introduce efficient $(1+varepsilon)$-approximation algorithms for the binary matrix factorization (BMF) problem.
The goal is to approximate $mathbfA$ as a product of low-rank factors.
Our techniques generalize to other common variants of the BMF problem.
arXiv Detail & Related papers (2023-06-02T18:55:27Z) - High-dimensional Location Estimation via Norm Concentration for Subgamma
Vectors [15.802475232604667]
In location estimation, we are given $n$ samples from a known distribution $f$ shifted by an unknown translation $lambda$.
Asymptotically, the maximum likelihood estimate achieves the Cram'er-Rao bound of error $mathcal N(0, frac1nmathcal I)$.
We build on the theory using emphsmoothed estimators to bound the error for finite $n$ in terms of $mathcal I_r$, the Fisher information of the $r$-smoothed
arXiv Detail & Related papers (2023-02-05T22:17:04Z) - TURF: A Two-factor, Universal, Robust, Fast Distribution Learning
Algorithm [64.13217062232874]
One of its most powerful and successful modalities approximates every distribution to an $ell$ distance essentially at most a constant times larger than its closest $t$-piece degree-$d_$.
We provide a method that estimates this number near-optimally, hence helps approach the best possible approximation.
arXiv Detail & Related papers (2022-02-15T03:49:28Z) - On Submodular Contextual Bandits [92.45432756301231]
We consider the problem of contextual bandits where actions are subsets of a ground set and mean rewards are modeled by an unknown monotone submodular function.
We show that our algorithm efficiently randomizes around local optima of estimated functions according to the Inverse Gap Weighting strategy.
arXiv Detail & Related papers (2021-12-03T21:42:33Z) - Non-Parametric Estimation of Manifolds from Noisy Data [1.0152838128195467]
We consider the problem of estimating a $d$ dimensional sub-manifold of $mathbbRD$ from a finite set of noisy samples.
We show that the estimation yields rates of convergence of $n-frack2k + d$ for the point estimation and $n-frack-12k + d$ for the estimation of tangent space.
arXiv Detail & Related papers (2021-05-11T02:29:33Z) - From Smooth Wasserstein Distance to Dual Sobolev Norm: Empirical
Approximation and Statistical Applications [18.618590805279187]
We show that $mathsfW_p(sigma)$ is controlled by a $pth order smooth dual Sobolev $mathsfd_p(sigma)$.
We derive the limit distribution of $sqrtnmathsfd_p(sigma)(hatmu_n,mu)$ in all dimensions $d$, when $mu$ is sub-Gaussian.
arXiv Detail & Related papers (2021-01-11T17:23:24Z) - Probabilistic Iterative Methods for Linear Systems [5.457842083043013]
We show that a standard iterative method on $mathbbRd$ is lifted to act on the space of probability distributions $mathcalP(mathbbRd)$.
The output of the iterative methods proposed in this paper is, instead, a sequence of probability distributions $mu_m in mathcalP(mathbbRd)$.
arXiv Detail & Related papers (2020-12-23T11:55:57Z) - Near-Optimal SQ Lower Bounds for Agnostically Learning Halfspaces and
ReLUs under Gaussian Marginals [49.60752558064027]
We study the fundamental problems of agnostically learning halfspaces and ReLUs under Gaussian marginals.
Our lower bounds provide strong evidence that current upper bounds for these tasks are essentially best possible.
arXiv Detail & Related papers (2020-06-29T17:10:10Z) - Agnostic Learning of a Single Neuron with Gradient Descent [92.7662890047311]
We consider the problem of learning the best-fitting single neuron as measured by the expected square loss.
For the ReLU activation, our population risk guarantee is $O(mathsfOPT1/2)+epsilon$.
For the ReLU activation, our population risk guarantee is $O(mathsfOPT1/2)+epsilon$.
arXiv Detail & Related papers (2020-05-29T07:20:35Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.