Learning and Covering Sums of Independent Random Variables with
Unbounded Support
- URL: http://arxiv.org/abs/2210.13313v1
- Date: Mon, 24 Oct 2022 15:03:55 GMT
- Title: Learning and Covering Sums of Independent Random Variables with
Unbounded Support
- Authors: Alkis Kalavasis, Konstantinos Stavropoulos, Manolis Zampetakis
- Abstract summary: We study the problem of covering and learning sums $X = X_1 + cdots + X_n$ of independent integer-valued random variables $X_i$ with unbounded, or even infinite, support.
We show that the maximum value of the collective support of $X_i$'s necessarily appears in the sample complexity of learning $X$.
- Score: 4.458210211781738
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study the problem of covering and learning sums $X = X_1 + \cdots + X_n$
of independent integer-valued random variables $X_i$ (SIIRVs) with unbounded,
or even infinite, support. De et al. at FOCS 2018, showed that the maximum
value of the collective support of $X_i$'s necessarily appears in the sample
complexity of learning $X$. In this work, we address two questions: (i) Are
there general families of SIIRVs with unbounded support that can be learned
with sample complexity independent of both $n$ and the maximal element of the
support? (ii) Are there general families of SIIRVs with unbounded support that
admit proper sparse covers in total variation distance? As for question (i), we
provide a set of simple conditions that allow the unbounded SIIRV to be learned
with complexity $\text{poly}(1/\epsilon)$ bypassing the aforementioned lower
bound. We further address question (ii) in the general setting where each
variable $X_i$ has unimodal probability mass function and is a different member
of some, possibly multi-parameter, exponential family $\mathcal{E}$ that
satisfies some structural properties. These properties allow $\mathcal{E}$ to
contain heavy tailed and non log-concave distributions. Moreover, we show that
for every $\epsilon > 0$, and every $k$-parameter family $\mathcal{E}$ that
satisfies some structural assumptions, there exists an algorithm with
$\tilde{O}(k) \cdot \text{poly}(1/\epsilon)$ samples that learns a sum of $n$
arbitrary members of $\mathcal{E}$ within $\epsilon$ in TV distance. The output
of the learning algorithm is also a sum of random variables whose distribution
lies in the family $\mathcal{E}$. En route, we prove that any discrete unimodal
exponential family with bounded constant-degree central moments can be
approximated by the family corresponding to a bounded subset of the initial
(unbounded) parameter space.
Related papers
- The Sample Complexity of Simple Binary Hypothesis Testing [7.127829790714167]
The sample complexity of simple binary hypothesis testing is the smallest number of i.i.d. samples required to distinguish between two distributions $p$ and $q$ in either setting.
This problem has only been studied when $alpha = beta$ (prior-free) or $alpha = 1/2$ (Bayesian)
arXiv Detail & Related papers (2024-03-25T17:42:32Z) - Outlier Robust Multivariate Polynomial Regression [27.03423421704806]
We are given a set of random samples $(mathbfx_i,y_i) in [-1,1]n times mathbbR$ that are noisy versions of $(mathbfx_i,p(mathbfx_i)$.
The goal is to output a $hatp$, within an $ell_in$-distance of at most $O(sigma)$ from $p$.
arXiv Detail & Related papers (2024-03-14T15:04:45Z) - Agnostically Learning Multi-index Models with Queries [54.290489524576756]
We study the power of query access for the task of agnostic learning under the Gaussian distribution.
We show that query access gives significant runtime improvements over random examples for agnostically learning MIMs.
arXiv Detail & Related papers (2023-12-27T15:50:47Z) - Testing Closeness of Multivariate Distributions via Ramsey Theory [40.926523210945064]
We investigate the statistical task of closeness (or equivalence) testing for multidimensional distributions.
Specifically, given sample access to two unknown distributions $mathbf p, mathbf q$ on $mathbb Rd$, we want to distinguish between the case that $mathbf p=mathbf q$ versus $|mathbf p-mathbf q|_A_k > epsilon$.
Our main result is the first closeness tester for this problem with em sub-learning sample complexity in any fixed dimension.
arXiv Detail & Related papers (2023-11-22T04:34:09Z) - Learning linear dynamical systems under convex constraints [4.4351901934764975]
We consider the problem of identification of linear dynamical systems from $T$ samples of a single trajectory.
$A*$ can be reliably estimated for values $T$ smaller than what is needed for unconstrained setting.
arXiv Detail & Related papers (2023-03-27T11:49:40Z) - Threshold Phenomena in Learning Halfspaces with Massart Noise [56.01192577666607]
We study the problem of PAC learning halfspaces on $mathbbRd$ with Massart noise under Gaussian marginals.
Our results qualitatively characterize the complexity of learning halfspaces in the Massart model.
arXiv Detail & Related papers (2021-08-19T16:16:48Z) - Small Covers for Near-Zero Sets of Polynomials and Learning Latent
Variable Models [56.98280399449707]
We show that there exists an $epsilon$-cover for $S$ of cardinality $M = (k/epsilon)O_d(k1/d)$.
Building on our structural result, we obtain significantly improved learning algorithms for several fundamental high-dimensional probabilistic models hidden variables.
arXiv Detail & Related papers (2020-12-14T18:14:08Z) - $k$-Forrelation Optimally Separates Quantum and Classical Query
Complexity [3.4984289152418753]
We show that any partial function on $N$ bits can be computed with an advantage $delta$ over a random guess by making $q$ quantum queries.
We also conjectured the $k$-Forrelation problem -- a partial function that can be computed with $q = lceil k/2 rceil$ quantum queries.
arXiv Detail & Related papers (2020-08-16T21:26:46Z) - Model-Free Reinforcement Learning: from Clipped Pseudo-Regret to Sample
Complexity [59.34067736545355]
Given an MDP with $S$ states, $A$ actions, the discount factor $gamma in (0,1)$, and an approximation threshold $epsilon > 0$, we provide a model-free algorithm to learn an $epsilon$-optimal policy.
For small enough $epsilon$, we show an improved algorithm with sample complexity.
arXiv Detail & Related papers (2020-06-06T13:34:41Z) - Agnostic Learning of a Single Neuron with Gradient Descent [92.7662890047311]
We consider the problem of learning the best-fitting single neuron as measured by the expected square loss.
For the ReLU activation, our population risk guarantee is $O(mathsfOPT1/2)+epsilon$.
For the ReLU activation, our population risk guarantee is $O(mathsfOPT1/2)+epsilon$.
arXiv Detail & Related papers (2020-05-29T07:20:35Z) - Locally Private Hypothesis Selection [96.06118559817057]
We output a distribution from $mathcalQ$ whose total variation distance to $p$ is comparable to the best such distribution.
We show that the constraint of local differential privacy incurs an exponential increase in cost.
Our algorithms result in exponential improvements on the round complexity of previous methods.
arXiv Detail & Related papers (2020-02-21T18:30:48Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.