DIFF2: Differential Private Optimization via Gradient Differences for
Nonconvex Distributed Learning
- URL: http://arxiv.org/abs/2302.03884v2
- Date: Sat, 3 Jun 2023 12:39:11 GMT
- Title: DIFF2: Differential Private Optimization via Gradient Differences for
Nonconvex Distributed Learning
- Authors: Tomoya Murata and Taiji Suzuki
- Abstract summary: In the previous work, the best known utility bound is $widetilde O(d2/3/(nvarepsilon_mathrmDP)4/3)$.
We propose a new differential private framework called mphDIFF2 (DIFFerential private via DIFFs) that constructs a differential private framework.
$mphDIFF2 with a global descent achieves the utility of $widetilde O(d2/3/(nvarepsilon_mathrmDP)4/3
- Score: 58.79085525115987
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Differential private optimization for nonconvex smooth objective is
considered. In the previous work, the best known utility bound is $\widetilde
O(\sqrt{d}/(n\varepsilon_\mathrm{DP}))$ in terms of the squared full gradient
norm, which is achieved by Differential Private Gradient Descent (DP-GD) as an
instance, where $n$ is the sample size, $d$ is the problem dimensionality and
$\varepsilon_\mathrm{DP}$ is the differential privacy parameter. To improve the
best known utility bound, we propose a new differential private optimization
framework called \emph{DIFF2 (DIFFerential private optimization via gradient
DIFFerences)} that constructs a differential private global gradient estimator
with possibly quite small variance based on communicated \emph{gradient
differences} rather than gradients themselves. It is shown that DIFF2 with a
gradient descent subroutine achieves the utility of $\widetilde
O(d^{2/3}/(n\varepsilon_\mathrm{DP})^{4/3})$, which can be significantly better
than the previous one in terms of the dependence on the sample size $n$. To the
best of our knowledge, this is the first fundamental result to improve the
standard utility $\widetilde O(\sqrt{d}/(n\varepsilon_\mathrm{DP}))$ for
nonconvex objectives. Additionally, a more computational and communication
efficient subroutine is combined with DIFF2 and its theoretical analysis is
also given. Numerical experiments are conducted to validate the superiority of
DIFF2 framework.
Related papers
- Differential Private Stochastic Optimization with Heavy-tailed Data: Towards Optimal Rates [15.27596975662702]
We explore algorithms achieving optimal rates of DP optimization with heavy-tailed gradients.
Our results match the minimax lower bound in citekamath2022, indicating that the theoretical limit of convex optimization under DP is achievable.
arXiv Detail & Related papers (2024-08-19T11:07:05Z) - PrivSGP-VR: Differentially Private Variance-Reduced Stochastic Gradient Push with Tight Utility Bounds [9.47030623916154]
We propose a differentially private decentralized learning method (termed PrivSGPVR) which employs gradient push with variance reduction and guarantees privacy for each node.
Our theoretical analysis shows that, under DP noise with constant variance, PrivGPS-VR achieves a sub-linear convergence rate of $mathcalO (1/sqrtnK)$.
arXiv Detail & Related papers (2024-05-04T11:22:53Z) - Mirror Descent Algorithms with Nearly Dimension-Independent Rates for
Differentially-Private Stochastic Saddle-Point Problems [6.431793114484429]
We propose $sqrtlog(d)/sqrtn + log(d)/[nvarepsilon]2/5$ to solve the problem of differentially-private saddle-points in the polyhedral setting.
We show that our algorithms attain a rate of $sqrtlog(d)/sqrtn + log(d)/[nvarepsilon]2/5$ with constant success.
arXiv Detail & Related papers (2024-03-05T12:28:00Z) - Variance-reduced Clipping for Non-convex Optimization [24.765794811146144]
Gradient clipping is a technique used in deep learning applications such as large-scale language modeling.
Recent experimental training have a fairly special behavior in that it mitigates order complexity.
arXiv Detail & Related papers (2023-03-02T00:57:38Z) - ReSQueing Parallel and Private Stochastic Convex Optimization [59.53297063174519]
We introduce a new tool for BFG convex optimization (SCO): a Reweighted Query (ReSQue) estimator for the gradient of a function convolved with a (Gaussian) probability density.
We develop algorithms achieving state-of-the-art complexities for SCO in parallel and private settings.
arXiv Detail & Related papers (2023-01-01T18:51:29Z) - Optimal Gradient Sliding and its Application to Distributed Optimization
Under Similarity [121.83085611327654]
We structured convex optimization problems with additive objective $r:=p + q$, where $r$ is $mu$-strong convex similarity.
We proposed a method to solve problems master to agents' communication and local calls.
The proposed method is much sharper than the $mathcalO(sqrtL_q/mu)$ method.
arXiv Detail & Related papers (2022-05-30T14:28:02Z) - Private Stochastic Convex Optimization: Optimal Rates in $\ell_1$
Geometry [69.24618367447101]
Up to logarithmic factors the optimal excess population loss of any $(varepsilon,delta)$-differently private is $sqrtlog(d)/n + sqrtd/varepsilon n.$
We show that when the loss functions satisfy additional smoothness assumptions, the excess loss is upper bounded (up to logarithmic factors) by $sqrtlog(d)/n + (log(d)/varepsilon n)2/3.
arXiv Detail & Related papers (2021-03-02T06:53:44Z) - Private Stochastic Non-Convex Optimization: Adaptive Algorithms and
Tighter Generalization Bounds [72.63031036770425]
We propose differentially private (DP) algorithms for bound non-dimensional optimization.
We demonstrate two popular deep learning methods on the empirical advantages over standard gradient methods.
arXiv Detail & Related papers (2020-06-24T06:01:24Z) - Maximizing Determinants under Matroid Constraints [69.25768526213689]
We study the problem of finding a basis $S$ of $M$ such that $det(sum_i in Sv_i v_i v_itop)$ is maximized.
This problem appears in a diverse set of areas such as experimental design, fair allocation of goods, network design, and machine learning.
arXiv Detail & Related papers (2020-04-16T19:16:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.