Differentially Private Algorithms for the Stochastic Saddle Point
Problem with Optimal Rates for the Strong Gap
- URL: http://arxiv.org/abs/2302.12909v2
- Date: Thu, 29 Jun 2023 16:22:34 GMT
- Title: Differentially Private Algorithms for the Stochastic Saddle Point
Problem with Optimal Rates for the Strong Gap
- Authors: Raef Bassily and Crist\'obal Guzm\'an and Michael Menart
- Abstract summary: We show that convex-concave Lipschitz saddle point problems can be solved under the constraint of $(epsilon,delta)$differential privacy.
We also show that there exists a fundamental tradeoff between stability and accuracy.
- Score: 12.446156563700482
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We show that convex-concave Lipschitz stochastic saddle point problems (also
known as stochastic minimax optimization) can be solved under the constraint of
$(\epsilon,\delta)$-differential privacy with \emph{strong (primal-dual) gap}
rate of $\tilde O\big(\frac{1}{\sqrt{n}} + \frac{\sqrt{d}}{n\epsilon}\big)$,
where $n$ is the dataset size and $d$ is the dimension of the problem. This
rate is nearly optimal, based on existing lower bounds in differentially
private stochastic optimization. Specifically, we prove a tight upper bound on
the strong gap via novel implementation and analysis of the recursive
regularization technique repurposed for saddle point problems. We show that
this rate can be attained with
$O\big(\min\big\{\frac{n^2\epsilon^{1.5}}{\sqrt{d}}, n^{3/2}\big\}\big)$
gradient complexity, and $\tilde{O}(n)$ gradient complexity if the loss
function is smooth. As a byproduct of our method, we develop a general
algorithm that, given a black-box access to a subroutine satisfying a certain
$\alpha$ primal-dual accuracy guarantee with respect to the empirical
objective, gives a solution to the stochastic saddle point problem with a
strong gap of $\tilde{O}(\alpha+\frac{1}{\sqrt{n}})$. We show that this
$\alpha$-accuracy condition is satisfied by standard algorithms for the
empirical saddle point problem such as the proximal point method and the
stochastic gradient descent ascent algorithm. Further, we show that even for
simple problems it is possible for an algorithm to have zero weak gap and
suffer from $\Omega(1)$ strong gap. We also show that there exists a
fundamental tradeoff between stability and accuracy. Specifically, we show that
any $\Delta$-stable algorithm has empirical gap $\Omega\big(\frac{1}{\Delta
n}\big)$, and that this bound is tight. This result also holds also more
specifically for empirical risk minimization problems and may be of independent
interest.
Related papers
- Mirror Descent Algorithms with Nearly Dimension-Independent Rates for
Differentially-Private Stochastic Saddle-Point Problems [6.431793114484429]
We propose $sqrtlog(d)/sqrtn + log(d)/[nvarepsilon]2/5$ to solve the problem of differentially-private saddle-points in the polyhedral setting.
We show that our algorithms attain a rate of $sqrtlog(d)/sqrtn + log(d)/[nvarepsilon]2/5$ with constant success.
arXiv Detail & Related papers (2024-03-05T12:28:00Z) - Differentially Private Non-Convex Optimization under the KL Condition with Optimal Rates [33.46648653842542]
We study private empirical risk problem for losses satisfying the $(gamma,kappa)$-Kurdyka-Lojasiewicz (KL) condition.
We show it is possible to achieve the rate $tildeObig(big(fracsqrtdnsqrtrhobig)kappabig)$ with a private implementation of the proximal point method.
arXiv Detail & Related papers (2023-11-22T15:12:42Z) - An Oblivious Stochastic Composite Optimization Algorithm for Eigenvalue
Optimization Problems [76.2042837251496]
We introduce two oblivious mirror descent algorithms based on a complementary composite setting.
Remarkably, both algorithms work without prior knowledge of the Lipschitz constant or smoothness of the objective function.
We show how to extend our framework to scale and demonstrate the efficiency and robustness of our methods on large scale semidefinite programs.
arXiv Detail & Related papers (2023-06-30T08:34:29Z) - Best Policy Identification in Linear MDPs [70.57916977441262]
We investigate the problem of best identification in discounted linear Markov+Delta Decision in the fixed confidence setting under a generative model.
The lower bound as the solution of an intricate non- optimization program can be used as the starting point to devise such algorithms.
arXiv Detail & Related papers (2022-08-11T04:12:50Z) - Lifted Primal-Dual Method for Bilinearly Coupled Smooth Minimax
Optimization [47.27237492375659]
We study the bilinearly coupled minimax problem: $min_x max_y f(x) + ytop A x - h(y)$, where $f$ and $h$ are both strongly convex smooth functions.
No known first-order algorithms have hitherto achieved the lower complexity bound of $Omega(sqrtfracL_xmu_x + frac|A|sqrtmu_x,mu_y) log(frac1vareps
arXiv Detail & Related papers (2022-01-19T05:56:19Z) - A first-order primal-dual method with adaptivity to local smoothness [64.62056765216386]
We consider the problem of finding a saddle point for the convex-concave objective $min_x max_y f(x) + langle Ax, yrangle - g*(y)$, where $f$ is a convex function with locally Lipschitz gradient and $g$ is convex and possibly non-smooth.
We propose an adaptive version of the Condat-Vu algorithm, which alternates between primal gradient steps and dual steps.
arXiv Detail & Related papers (2021-10-28T14:19:30Z) - No quantum speedup over gradient descent for non-smooth convex
optimization [22.16973542453584]
Black-box access to a (not necessarily smooth) function $f:mathbbRn to mathbbR$ and its (sub)gradient.
Our goal is to find an $epsilon$-approximate minimum of $f$ starting from a point that is distance at most $R$ from the true minimum.
We show that although the function family used in the lower bound is hard for randomized algorithms, it can be solved using $O(GR/epsilon)$ quantum queries.
arXiv Detail & Related papers (2020-10-05T06:32:47Z) - Hybrid Stochastic-Deterministic Minibatch Proximal Gradient:
Less-Than-Single-Pass Optimization with Nearly Optimal Generalization [83.80460802169999]
We show that HSDMPG can attain an $mathcalObig (1/sttnbig)$ which is at the order of excess error on a learning model.
For loss factors, we prove that HSDMPG can attain an $mathcalObig (1/sttnbig)$ which is at the order of excess error on a learning model.
arXiv Detail & Related papers (2020-09-18T02:18:44Z) - Streaming Complexity of SVMs [110.63976030971106]
We study the space complexity of solving the bias-regularized SVM problem in the streaming model.
We show that for both problems, for dimensions of $frac1lambdaepsilon$, one can obtain streaming algorithms with spacely smaller than $frac1lambdaepsilon$.
arXiv Detail & Related papers (2020-07-07T17:10:00Z) - Maximizing Determinants under Matroid Constraints [69.25768526213689]
We study the problem of finding a basis $S$ of $M$ such that $det(sum_i in Sv_i v_i v_itop)$ is maximized.
This problem appears in a diverse set of areas such as experimental design, fair allocation of goods, network design, and machine learning.
arXiv Detail & Related papers (2020-04-16T19:16:38Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.