Efficient Distance Approximation for Structured High-Dimensional
Distributions via Learning
- URL: http://arxiv.org/abs/2002.05378v2
- Date: Fri, 14 Feb 2020 03:03:10 GMT
- Title: Efficient Distance Approximation for Structured High-Dimensional
Distributions via Learning
- Authors: Arnab Bhattacharyya, Sutanu Gayen, Kuldeep S. Meel, N. V.
Vinodchandran
- Abstract summary: We design efficient distance approximation algorithms for several classes of structured high-dimensional distributions.
Our results are the first efficient distance approximation algorithms for these well-studied problems.
- Score: 31.552581550603005
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We design efficient distance approximation algorithms for several classes of
structured high-dimensional distributions. Specifically, we show algorithms for
the following problems:
- Given sample access to two Bayesian networks $P_1$ and $P_2$ over known
directed acyclic graphs $G_1$ and $G_2$ having $n$ nodes and bounded in-degree,
approximate $d_{tv}(P_1,P_2)$ to within additive error $\epsilon$ using
$poly(n,\epsilon)$ samples and time
- Given sample access to two ferromagnetic Ising models $P_1$ and $P_2$ on
$n$ variables with bounded width, approximate $d_{tv}(P_1, P_2)$ to within
additive error $\epsilon$ using $poly(n,\epsilon)$ samples and time
- Given sample access to two $n$-dimensional Gaussians $P_1$ and $P_2$,
approximate $d_{tv}(P_1, P_2)$ to within additive error $\epsilon$ using
$poly(n,\epsilon)$ samples and time
- Given access to observations from two causal models $P$ and $Q$ on $n$
variables that are defined over known causal graphs, approximate $d_{tv}(P_a,
Q_a)$ to within additive error $\epsilon$ using $poly(n,\epsilon)$ samples,
where $P_a$ and $Q_a$ are the interventional distributions obtained by the
intervention $do(A=a)$ on $P$ and $Q$ respectively for a particular variable
$A$.
Our results are the first efficient distance approximation algorithms for
these well-studied problems. They are derived using a simple and general
connection to distribution learning algorithms. The distance approximation
algorithms imply new efficient algorithms for {\em tolerant} testing of
closeness of the above-mentioned structured high-dimensional distributions.
Related papers
- Robust Distribution Learning with Local and Global Adversarial Corruptions [17.22168727622332]
We develop an efficient finite-sample algorithm with error bounded by $sqrtvarepsilon k + rho + tildeO(dsqrtkn-1/(k lor 2))$ when $P$ has bounded covariance.
Our efficient procedure relies on a novel trace norm approximation of an ideal yet intractable 2-Wasserstein projection estimator.
arXiv Detail & Related papers (2024-06-10T17:48:36Z) - Finding good policies in average-reward Markov Decision Processes without prior knowledge [19.89784209009327]
We revisit the identification of an $varepsilon$-optimal policy in average-reward Decision (MDP)
By relying on a diameter estimation procedure, we propose the first algorithm for $(varepsilon,delta)$-PAC-PAC policy identification.
arXiv Detail & Related papers (2024-05-27T12:24:14Z) - Quantum (Inspired) $D^2$-sampling with Applications [0.138120109831448]
We show a quantum implementation of $k$-means++ that runs in time $tildeO(zeta2 k2)$.
It can be shown through a robust approximation analysis of $k$-means++ that the quantum version preserves its $O(logk)$ approximation guarantee.
This results in a fast quantum-inspired classical implementation of $k$-means++, which we call QI-$k$-means++, with a running time $O(Nd) + tildeO(zeta
arXiv Detail & Related papers (2024-05-22T05:13:28Z) - Data Structures for Density Estimation [66.36971978162461]
Given a sublinear (in $n$) number of samples from $p$, our main result is the first data structure that identifies $v_i$ in time sublinear in $k$.
We also give an improved version of the algorithm of Acharya et al. that reports $v_i$ in time linear in $k$.
arXiv Detail & Related papers (2023-06-20T06:13:56Z) - Beyond Moments: Robustly Learning Affine Transformations with
Asymptotically Optimal Error [8.615625517708324]
We present a-time algorithm for learning an unknown affine transformation of the standard hypercube from samples.
Our algorithm is based on a new method that iteratively improves an estimate of the unknown affine transformation whenever the requirements of the certificate are not met.
arXiv Detail & Related papers (2023-02-23T19:13:30Z) - TURF: A Two-factor, Universal, Robust, Fast Distribution Learning
Algorithm [64.13217062232874]
One of its most powerful and successful modalities approximates every distribution to an $ell$ distance essentially at most a constant times larger than its closest $t$-piece degree-$d_$.
We provide a method that estimates this number near-optimally, hence helps approach the best possible approximation.
arXiv Detail & Related papers (2022-02-15T03:49:28Z) - Improved Sample Complexity for Incremental Autonomous Exploration in
MDPs [132.88757893161699]
We learn the set of $epsilon$-optimal goal-conditioned policies attaining all states that are incrementally reachable within $L$ steps.
DisCo is the first algorithm that can return an $epsilon/c_min$-optimal policy for any cost-sensitive shortest-path problem.
arXiv Detail & Related papers (2020-12-29T14:06:09Z) - Small Covers for Near-Zero Sets of Polynomials and Learning Latent
Variable Models [56.98280399449707]
We show that there exists an $epsilon$-cover for $S$ of cardinality $M = (k/epsilon)O_d(k1/d)$.
Building on our structural result, we obtain significantly improved learning algorithms for several fundamental high-dimensional probabilistic models hidden variables.
arXiv Detail & Related papers (2020-12-14T18:14:08Z) - Optimal Robust Linear Regression in Nearly Linear Time [97.11565882347772]
We study the problem of high-dimensional robust linear regression where a learner is given access to $n$ samples from the generative model $Y = langle X,w* rangle + epsilon$
We propose estimators for this problem under two settings: (i) $X$ is L4-L2 hypercontractive, $mathbbE [XXtop]$ has bounded condition number and $epsilon$ has bounded variance and (ii) $X$ is sub-Gaussian with identity second moment and $epsilon$ is
arXiv Detail & Related papers (2020-07-16T06:44:44Z) - Model-Free Reinforcement Learning: from Clipped Pseudo-Regret to Sample
Complexity [59.34067736545355]
Given an MDP with $S$ states, $A$ actions, the discount factor $gamma in (0,1)$, and an approximation threshold $epsilon > 0$, we provide a model-free algorithm to learn an $epsilon$-optimal policy.
For small enough $epsilon$, we show an improved algorithm with sample complexity.
arXiv Detail & Related papers (2020-06-06T13:34:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.