On The Convergence of First Order Methods for Quasar-Convex Optimization
- URL: http://arxiv.org/abs/2010.04937v3
- Date: Tue, 27 Oct 2020 11:58:31 GMT
- Title: On The Convergence of First Order Methods for Quasar-Convex Optimization
- Authors: Jikai Jin
- Abstract summary: In recent years, the success of deep learning has inspired many researchers to study general smooth non-sarsar functions.
In this paper, we study the convergence of first methods in a variety of different different settings.
- Score: 1.52292571922932
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, the success of deep learning has inspired many researchers
to study the optimization of general smooth non-convex functions. However,
recent works have established pessimistic worst-case complexities for this
class functions, which is in stark contrast with their superior performance in
real-world applications (e.g. training deep neural networks). On the other
hand, it is found that many popular non-convex optimization problems enjoy
certain structured properties which bear some similarities to convexity. In
this paper, we study the class of \textit{quasar-convex functions} to close the
gap between theory and practice. We study the convergence of first order
methods in a variety of different settings and under different optimality
criterions. We prove complexity upper bounds that are similar to standard
results established for convex functions and much better that state-of-the-art
convergence rates of non-convex functions. Overall, this paper suggests that
\textit{quasar-convexity} allows efficient optimization procedures, and we are
looking forward to seeing more problems that demonstrate similar properties in
practice.
Related papers
- Linearization Algorithms for Fully Composite Optimization [61.20539085730636]
This paper studies first-order algorithms for solving fully composite optimization problems convex compact sets.
We leverage the structure of the objective by handling differentiable and non-differentiable separately, linearizing only the smooth parts.
arXiv Detail & Related papers (2023-02-24T18:41:48Z) - High-Probability Bounds for Stochastic Optimization and Variational
Inequalities: the Case of Unbounded Variance [59.211456992422136]
We propose algorithms with high-probability convergence results under less restrictive assumptions.
These results justify the usage of the considered methods for solving problems that do not fit standard functional classes in optimization.
arXiv Detail & Related papers (2023-02-02T10:37:23Z) - Efficient First-order Methods for Convex Optimization with Strongly
Convex Function Constraints [3.667453772837954]
We show how to minimize a convex function subject to strongly convex function constraints.
We identify the sparsity pattern within a finite number result that appears to have independent significance.
arXiv Detail & Related papers (2022-12-21T16:04:53Z) - Lower Bounds and Accelerated Algorithms for Bilevel Optimization [62.192297758346484]
Bilevel optimization has recently attracted growing interests due to its wide applications in modern machine learning problems.
We show that our results are more challenging than that of minimax applications.
arXiv Detail & Related papers (2021-02-07T21:46:29Z) - Recent Theoretical Advances in Non-Convex Optimization [56.88981258425256]
Motivated by recent increased interest in analysis of optimization algorithms for non- optimization in deep networks and other problems in data, we give an overview of recent results of theoretical optimization algorithms for non- optimization.
arXiv Detail & Related papers (2020-12-11T08:28:51Z) - Efficient Methods for Structured Nonconvex-Nonconcave Min-Max
Optimization [98.0595480384208]
We propose a generalization extraient spaces which converges to a stationary point.
The algorithm applies not only to general $p$-normed spaces, but also to general $p$-dimensional vector spaces.
arXiv Detail & Related papers (2020-10-31T21:35:42Z) - Fast Objective & Duality Gap Convergence for Non-Convex Strongly-Concave
Min-Max Problems with PL Condition [52.08417569774822]
This paper focuses on methods for solving smooth non-concave min-max problems, which have received increasing attention due to deep learning (e.g., deep AUC)
arXiv Detail & Related papers (2020-06-12T00:32:21Z) - Do optimization methods in deep learning applications matter? [0.0]
The paper presents arguments on which optimization functions to use and further, which functions would benefit from parallelization efforts.
Our experiments compare off-the-shelf optimization functions(CG, SGD, LM and L-BFGS) in standard CIFAR, MNIST, CartPole and FlappyBird experiments.
arXiv Detail & Related papers (2020-02-28T10:36:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.