Private Stochastic Convex Optimization: Optimal Rates in Linear Time
- URL: http://arxiv.org/abs/2005.04763v1
- Date: Sun, 10 May 2020 19:52:03 GMT
- Title: Private Stochastic Convex Optimization: Optimal Rates in Linear Time
- Authors: Vitaly Feldman, Tomer Koren, Kunal Talwar
- Abstract summary: We study the problem of minimizing the population loss given i.i.d. samples from a distribution over convex loss functions.
A recent work of Bassily et al. has established the optimal bound on the excess population loss achievable given $n$ samples.
We describe two new techniques for deriving convex optimization algorithms both achieving the optimal bound on excess loss and using $O(minn, n2/d)$ gradient computations.
- Score: 74.47681868973598
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study differentially private (DP) algorithms for stochastic convex
optimization: the problem of minimizing the population loss given i.i.d.
samples from a distribution over convex loss functions. A recent work of
Bassily et al. (2019) has established the optimal bound on the excess
population loss achievable given $n$ samples. Unfortunately, their algorithm
achieving this bound is relatively inefficient: it requires $O(\min\{n^{3/2},
n^{5/2}/d\})$ gradient computations, where $d$ is the dimension of the
optimization problem.
We describe two new techniques for deriving DP convex optimization algorithms
both achieving the optimal bound on excess loss and using $O(\min\{n, n^2/d\})$
gradient computations. In particular, the algorithms match the running time of
the optimal non-private algorithms. The first approach relies on the use of
variable batch sizes and is analyzed using the privacy amplification by
iteration technique of Feldman et al. (2018). The second approach is based on a
general reduction to the problem of localizing an approximately optimal
solution with differential privacy. Such localization, in turn, can be achieved
using existing (non-private) uniformly stable optimization algorithms. As in
the earlier work, our algorithms require a mild smoothness assumption. We also
give a linear-time algorithm achieving the optimal bound on the excess loss for
the strongly convex case, as well as a faster algorithm for the non-smooth
case.
Related papers
- Accelerating Cutting-Plane Algorithms via Reinforcement Learning
Surrogates [49.84541884653309]
A current standard approach to solving convex discrete optimization problems is the use of cutting-plane algorithms.
Despite the existence of a number of general-purpose cut-generating algorithms, large-scale discrete optimization problems continue to suffer from intractability.
We propose a method for accelerating cutting-plane algorithms via reinforcement learning.
arXiv Detail & Related papers (2023-07-17T20:11:56Z) - An Algorithm with Optimal Dimension-Dependence for Zero-Order Nonsmooth Nonconvex Stochastic Optimization [37.300102993926046]
We study the complexity of producing neither smooth nor convex points of Lipschitz objectives which are possibly using only zero-order evaluations.
Our analysis is based on a simple yet powerful.
Goldstein-subdifferential set, which allows recent advancements in.
nonsmooth non optimization.
arXiv Detail & Related papers (2023-07-10T11:56:04Z) - Accelerated First-Order Optimization under Nonlinear Constraints [73.2273449996098]
We exploit between first-order algorithms for constrained optimization and non-smooth systems to design a new class of accelerated first-order algorithms.
An important property of these algorithms is that constraints are expressed in terms of velocities instead of sparse variables.
arXiv Detail & Related papers (2023-02-01T08:50:48Z) - Fast Computation of Optimal Transport via Entropy-Regularized Extragradient Methods [75.34939761152587]
Efficient computation of the optimal transport distance between two distributions serves as an algorithm that empowers various applications.
This paper develops a scalable first-order optimization-based method that computes optimal transport to within $varepsilon$ additive accuracy.
arXiv Detail & Related papers (2023-01-30T15:46:39Z) - Projected Stochastic Gradient Langevin Algorithms for Constrained
Sampling and Non-Convex Learning [0.0]
Langevin algorithms are methods with additive noise.
Langevin algorithms have been used for decades in chain Carlo (Milon)
For learning, we show that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that it is that
arXiv Detail & Related papers (2020-12-22T16:19:20Z) - A Two-Timescale Framework for Bilevel Optimization: Complexity Analysis
and Application to Actor-Critic [142.1492359556374]
Bilevel optimization is a class of problems which exhibit a two-level structure.
We propose a two-timescale approximation (TTSA) algorithm for tackling such a bilevel problem.
We show that a two-timescale natural actor-critic policy optimization algorithm can be viewed as a special case of our TTSA framework.
arXiv Detail & Related papers (2020-07-10T05:20:02Z) - Convergence of adaptive algorithms for weakly convex constrained
optimization [59.36386973876765]
We prove the $mathcaltilde O(t-1/4)$ rate of convergence for the norm of the gradient of Moreau envelope.
Our analysis works with mini-batch size of $1$, constant first and second order moment parameters, and possibly smooth optimization domains.
arXiv Detail & Related papers (2020-06-11T17:43:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.