The Complexity of Gradient Descent: CLS = PPAD $\cap$ PLS
- URL: http://arxiv.org/abs/2011.01929v3
- Date: Tue, 13 Apr 2021 14:51:12 GMT
- Title: The Complexity of Gradient Descent: CLS = PPAD $\cap$ PLS
- Authors: John Fearnley, Paul W. Goldberg, Alexandros Hollender, Rahul Savani
- Abstract summary: We study search problems that can be solved by performing Gradient Descent on a bounded convex polytopal domain.
We show that this class is equal to the intersection of two well-known classes: PPAD and PLS.
- Score: 68.419966284392
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We study search problems that can be solved by performing Gradient Descent on
a bounded convex polytopal domain and show that this class is equal to the
intersection of two well-known classes: PPAD and PLS. As our main underlying
technical contribution, we show that computing a Karush-Kuhn-Tucker (KKT) point
of a continuously differentiable function over the domain $[0,1]^2$ is PPAD
$\cap$ PLS-complete. This is the first natural problem to be shown complete for
this class. Our results also imply that the class CLS (Continuous Local Search)
- which was defined by Daskalakis and Papadimitriou as a more "natural"
counterpart to PPAD $\cap$ PLS and contains many interesting problems - is
itself equal to PPAD $\cap$ PLS.
Related papers
- Single-Loop Stochastic Algorithms for Difference of Max-Structured Weakly Convex Functions [41.43895948769255]
We show a class of non-smooth non-asymptotic fairness problems in the form of $min_x[yin Yphi(x, y) - max_zin Zpsix(x, z)]$.
We propose an envelope approximate gradient SMAG, the first method for solving these problems, provide a state-of-the-art non-asymptotic convergence rate.
arXiv Detail & Related papers (2024-05-28T20:52:46Z) - On the complexity of PAC learning in Hilbert spaces [0.0]
We study the problem of binary classification from the point of view of learning convex polyhedra in Hilbert spaces.
We propose an algorithm for learning a polyhedron which correctly classifies at least $1- varepsilon$ of the distribution.
arXiv Detail & Related papers (2023-03-03T16:16:11Z) - Detection-Recovery Gap for Planted Dense Cycles [72.4451045270967]
We consider a model where a dense cycle with expected bandwidth $n tau$ and edge density $p$ is planted in an ErdHos-R'enyi graph $G(n,q)$.
We characterize the computational thresholds for the associated detection and recovery problems for the class of low-degree algorithms.
arXiv Detail & Related papers (2023-02-13T22:51:07Z) - Optimal Gradient Sliding and its Application to Distributed Optimization
Under Similarity [121.83085611327654]
We structured convex optimization problems with additive objective $r:=p + q$, where $r$ is $mu$-strong convex similarity.
We proposed a method to solve problems master to agents' communication and local calls.
The proposed method is much sharper than the $mathcalO(sqrtL_q/mu)$ method.
arXiv Detail & Related papers (2022-05-30T14:28:02Z) - Learning Stochastic Shortest Path with Linear Function Approximation [74.08819218747341]
We study the shortest path (SSP) problem in reinforcement learning with linear function approximation, where the transition kernel is represented as a linear mixture of unknown models.
We propose a novel algorithm for learning the linear mixture SSP, which can attain a $tilde O(d B_star1.5sqrtK/c_min)$ regret.
arXiv Detail & Related papers (2021-10-25T08:34:00Z) - On Agnostic PAC Learning using $\mathcal{L}_2$-polynomial Regression and
Fourier-based Algorithms [10.66048003460524]
We develop a framework using Hilbert spaces as a proxy to analyze PAC learning problems with structural properties.
We demonstrate that PAC learning with 0-1 loss is equivalent to an optimization in the Hilbert space domain.
arXiv Detail & Related papers (2021-02-11T21:28:55Z) - Small Covers for Near-Zero Sets of Polynomials and Learning Latent
Variable Models [56.98280399449707]
We show that there exists an $epsilon$-cover for $S$ of cardinality $M = (k/epsilon)O_d(k1/d)$.
Building on our structural result, we obtain significantly improved learning algorithms for several fundamental high-dimensional probabilistic models hidden variables.
arXiv Detail & Related papers (2020-12-14T18:14:08Z) - Algorithms and SQ Lower Bounds for PAC Learning One-Hidden-Layer ReLU
Networks [48.32532049640782]
We study the problem of learning one-hidden-layer ReLU networks with $k$ hidden units on $mathbbRd$ under Gaussian marginals.
For the case of positive coefficients, we give the first-time algorithm for this learning problem for $k$ up to $tildeOOmega(sqrtlog d)$.
arXiv Detail & Related papers (2020-06-22T17:53:54Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.