\~Optimal Differentially Private Learning of Thresholds and
Quasi-Concave Optimization
- URL: http://arxiv.org/abs/2211.06387v1
- Date: Fri, 11 Nov 2022 18:16:46 GMT
- Title: \~Optimal Differentially Private Learning of Thresholds and
Quasi-Concave Optimization
- Authors: Edith Cohen, Xin Lyu, Jelani Nelson, Tam\'as Sarl\'os, Uri Stemmer
- Abstract summary: The problem of learning threshold functions is a fundamental one in machine learning.
We provide a nearly tight upper bound of $tildeO(log* |X|)1.5)$ by Kaplan et al.
We also provide matching upper and lower bounds of $tildeTheta(2log*|X|)$ for the additive error of private quasi-concave (a related and more general problem)
- Score: 23.547605262139957
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: The problem of learning threshold functions is a fundamental one in machine
learning. Classical learning theory implies sample complexity of $O(\xi^{-1}
\log(1/\beta))$ (for generalization error $\xi$ with confidence $1-\beta$). The
private version of the problem, however, is more challenging and in particular,
the sample complexity must depend on the size $|X|$ of the domain. Progress on
quantifying this dependence, via lower and upper bounds, was made in a line of
works over the past decade. In this paper, we finally close the gap for
approximate-DP and provide a nearly tight upper bound of $\tilde{O}(\log^*
|X|)$, which matches a lower bound by Alon et al (that applies even with
improper learning) and improves over a prior upper bound of $\tilde{O}((\log^*
|X|)^{1.5})$ by Kaplan et al. We also provide matching upper and lower bounds
of $\tilde{\Theta}(2^{\log^*|X|})$ for the additive error of private
quasi-concave optimization (a related and more general problem). Our
improvement is achieved via the novel Reorder-Slice-Compute paradigm for
private data analysis which we believe will have further applications.
Related papers
- Private Isotonic Regression [54.32252900997422]
We consider the problem of isotonic regression over a partially ordered set (poset) $mathcalX$ and for any Lipschitz loss function.
We obtain a pure-DP algorithm that has an expected excess empirical risk of roughly $mathrmwidth(mathcalX) cdot log|mathcalX| / n$, where $mathrmwidth(mathcalX)$ is the width of the poset.
We show that the bounds above are essentially the best that can be
arXiv Detail & Related papers (2022-10-27T05:08:07Z) - Optimal Query Complexities for Dynamic Trace Estimation [59.032228008383484]
We consider the problem of minimizing the number of matrix-vector queries needed for accurate trace estimation in the dynamic setting where our underlying matrix is changing slowly.
We provide a novel binary tree summation procedure that simultaneously estimates all $m$ traces up to $epsilon$ error with $delta$ failure probability.
Our lower bounds (1) give the first tight bounds for Hutchinson's estimator in the matrix-vector product model with Frobenius norm error even in the static setting, and (2) are the first unconditional lower bounds for dynamic trace estimation.
arXiv Detail & Related papers (2022-09-30T04:15:44Z) - On the Sample Complexity of Privately Learning Axis-Aligned Rectangles [16.092248433189816]
We revisit the fundamental problem of learning Axis-Aligned-Rectangles over a finite grid.
We present a novel algorithm that reduces the sample complexity to only $tildeOleftdcdotleft(log*|X|right)1.5right$.
arXiv Detail & Related papers (2021-07-24T04:06:11Z) - Private Stochastic Convex Optimization: Optimal Rates in $\ell_1$
Geometry [69.24618367447101]
Up to logarithmic factors the optimal excess population loss of any $(varepsilon,delta)$-differently private is $sqrtlog(d)/n + sqrtd/varepsilon n.$
We show that when the loss functions satisfy additional smoothness assumptions, the excess loss is upper bounded (up to logarithmic factors) by $sqrtlog(d)/n + (log(d)/varepsilon n)2/3.
arXiv Detail & Related papers (2021-03-02T06:53:44Z) - Hybrid Stochastic-Deterministic Minibatch Proximal Gradient:
Less-Than-Single-Pass Optimization with Nearly Optimal Generalization [83.80460802169999]
We show that HSDMPG can attain an $mathcalObig (1/sttnbig)$ which is at the order of excess error on a learning model.
For loss factors, we prove that HSDMPG can attain an $mathcalObig (1/sttnbig)$ which is at the order of excess error on a learning model.
arXiv Detail & Related papers (2020-09-18T02:18:44Z) - Model-Free Reinforcement Learning: from Clipped Pseudo-Regret to Sample
Complexity [59.34067736545355]
Given an MDP with $S$ states, $A$ actions, the discount factor $gamma in (0,1)$, and an approximation threshold $epsilon > 0$, we provide a model-free algorithm to learn an $epsilon$-optimal policy.
For small enough $epsilon$, we show an improved algorithm with sample complexity.
arXiv Detail & Related papers (2020-06-06T13:34:41Z) - Agnostic Q-learning with Function Approximation in Deterministic
Systems: Tight Bounds on Approximation Error and Sample Complexity [94.37110094442136]
We study the problem of agnostic $Q$-learning with function approximation in deterministic systems.
We show that if $delta = Oleft(rho/sqrtdim_Eright)$, then one can find the optimal policy using $Oleft(dim_Eright)$.
arXiv Detail & Related papers (2020-02-17T18:41:49Z) - Differentially Private Release and Learning of Threshold Functions [27.612916837481485]
We prove new upper and lower bounds on the sample complexity of $(epsilon, delta)$ differentially private algorithms.
A threshold function $c_x$ over a totally ordered domain $X$ evaluates to $c_x(y) = 1$ if $y le x$, and evaluates to $0$ otherwise.
arXiv Detail & Related papers (2015-04-28T16:15:01Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.