On Differentially Private Stochastic Convex Optimization with
Heavy-tailed Data
- URL: http://arxiv.org/abs/2010.11082v1
- Date: Wed, 21 Oct 2020 15:45:27 GMT
- Title: On Differentially Private Stochastic Convex Optimization with
Heavy-tailed Data
- Authors: Di Wang and Hanshen Xiao and Srini Devadas and Jinhui Xu
- Abstract summary: We consider the problem of designing Differentially Private (DP) algorithms for Convex Optimization (SCO) on heavy-tailed data.
The irregularity of such data violates some key assumptions used in almost all existing DP-SCO and DP-ERM methods.
We show that our algorithms can effectively deal with the challenges caused by data irregularity.
- Score: 18.351352536791453
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we consider the problem of designing Differentially Private
(DP) algorithms for Stochastic Convex Optimization (SCO) on heavy-tailed data.
The irregularity of such data violates some key assumptions used in almost all
existing DP-SCO and DP-ERM methods, resulting in failure to provide the DP
guarantees. To better understand this type of challenges, we provide in this
paper a comprehensive study of DP-SCO under various settings. First, we
consider the case where the loss function is strongly convex and smooth. For
this case, we propose a method based on the sample-and-aggregate framework,
which has an excess population risk of $\tilde{O}(\frac{d^3}{n\epsilon^4})$
(after omitting other factors), where $n$ is the sample size and $d$ is the
dimensionality of the data. Then, we show that with some additional assumptions
on the loss functions, it is possible to reduce the \textit{expected} excess
population risk to $\tilde{O}(\frac{ d^2}{ n\epsilon^2 })$. To lift these
additional conditions, we also provide a gradient smoothing and trimming based
scheme to achieve excess population risks of $\tilde{O}(\frac{
d^2}{n\epsilon^2})$ and
$\tilde{O}(\frac{d^\frac{2}{3}}{(n\epsilon^2)^\frac{1}{3}})$ for strongly
convex and general convex loss functions, respectively, \textit{with high
probability}. Experiments suggest that our algorithms can effectively deal with
the challenges caused by data irregularity.
Related papers
- Inverse Entropic Optimal Transport Solves Semi-supervised Learning via Data Likelihood Maximization [65.8915778873691]
conditional distributions is a central problem in machine learning.
We propose a new learning paradigm that integrates both paired and unpaired data.
Our approach also connects intriguingly with inverse entropic optimal transport (OT)
arXiv Detail & Related papers (2024-10-03T16:12:59Z) - Corruption-Robust Offline Reinforcement Learning with General Function
Approximation [60.91257031278004]
We investigate the problem of corruption in offline reinforcement learning (RL) with general function approximation.
Our goal is to find a policy that is robust to such corruption and minimizes the suboptimality gap with respect to the optimal policy for the uncorrupted Markov decision processes (MDPs)
arXiv Detail & Related papers (2023-10-23T04:07:26Z) - Differentially Private Stochastic Convex Optimization in (Non)-Euclidean
Space Revisited [6.339471679859413]
We revisit the problem of Differentially Private Convex (DP-SCO) in Euclidelldd space.
For both strongly bound and strongly bound loss functions, we could achieve () outputs that are only dependent on the excess population.
We also show the width of the convex functions is optimal up to a logarithmic factor.
arXiv Detail & Related papers (2023-03-31T13:29:27Z) - ReSQueing Parallel and Private Stochastic Convex Optimization [59.53297063174519]
We introduce a new tool for BFG convex optimization (SCO): a Reweighted Query (ReSQue) estimator for the gradient of a function convolved with a (Gaussian) probability density.
We develop algorithms achieving state-of-the-art complexities for SCO in parallel and private settings.
arXiv Detail & Related papers (2023-01-01T18:51:29Z) - Normalized/Clipped SGD with Perturbation for Differentially Private
Non-Convex Optimization [94.06564567766475]
DP-SGD and DP-NSGD mitigate the risk of large models memorizing sensitive training data.
We show that these two algorithms achieve similar best accuracy while DP-NSGD is comparatively easier to tune than DP-SGD.
arXiv Detail & Related papers (2022-06-27T03:45:02Z) - Sharper Utility Bounds for Differentially Private Models [20.246768861331276]
First $mathcalObig(fracsqrtpnepsilonbig)$ high probability excess population risk bound for differentially private algorithms.
New proposed algorithm m-NGP improves the performance of the differentially private model over real datasets.
arXiv Detail & Related papers (2022-04-22T07:03:13Z) - Settling the Sample Complexity of Model-Based Offline Reinforcement
Learning [50.5790774201146]
offline reinforcement learning (RL) learns using pre-collected data without further exploration.
Prior algorithms or analyses either suffer from suboptimal sample complexities or incur high burn-in cost to reach sample optimality.
We demonstrate that the model-based (or "plug-in") approach achieves minimax-optimal sample complexity without burn-in cost.
arXiv Detail & Related papers (2022-04-11T17:26:19Z) - High Dimensional Differentially Private Stochastic Optimization with
Heavy-tailed Data [8.55881355051688]
We provide the first study on the problem of DP-SCO with heavy-tailed data in the high dimensional space.
We show that if the loss function is smooth and its gradient has bounded second order moment, it is possible to get a (high probability) error bound (excess population risk) of $tildeO(fraclog d(nepsilon)frac13)$ in the $epsilon$-DP model.
In the second part of the paper, we study sparse learning with heavy-tailed data.
arXiv Detail & Related papers (2021-07-23T11:03:21Z) - Distributionally Robust Optimization with Markovian Data [8.126833795693699]
We study a program where the probability distribution of the uncertain problem parameters is unknown.
We propose a data-driven distributionally to estimate the problem's objective function and optimal solution.
arXiv Detail & Related papers (2021-06-12T10:59:02Z) - Non-Euclidean Differentially Private Stochastic Convex Optimization [15.302167005107135]
We show that noisy gradient descent (SGD) algorithms attain the optimal excess risk in low-dimensional regimes.
Our work draws upon concepts from the geometry of normed spaces, such as the notions of regularity, uniform convexity, and uniform smoothness.
arXiv Detail & Related papers (2021-03-01T19:48:44Z) - Private Stochastic Non-Convex Optimization: Adaptive Algorithms and
Tighter Generalization Bounds [72.63031036770425]
We propose differentially private (DP) algorithms for bound non-dimensional optimization.
We demonstrate two popular deep learning methods on the empirical advantages over standard gradient methods.
arXiv Detail & Related papers (2020-06-24T06:01:24Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.