On The Convergence of First Order Methods for Quasar-Convex Optimization
- URL: http://arxiv.org/abs/2010.04937v3
- Date: Tue, 27 Oct 2020 11:58:31 GMT
- Title: On The Convergence of First Order Methods for Quasar-Convex Optimization
- Authors: Jikai Jin
- Abstract summary: In recent years, the success of deep learning has inspired many researchers to study general smooth non-sarsar functions.
In this paper, we study the convergence of first methods in a variety of different different settings.
- Score: 1.52292571922932
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In recent years, the success of deep learning has inspired many researchers
to study the optimization of general smooth non-convex functions. However,
recent works have established pessimistic worst-case complexities for this
class functions, which is in stark contrast with their superior performance in
real-world applications (e.g. training deep neural networks). On the other
hand, it is found that many popular non-convex optimization problems enjoy
certain structured properties which bear some similarities to convexity. In
this paper, we study the class of \textit{quasar-convex functions} to close the
gap between theory and practice. We study the convergence of first order
methods in a variety of different settings and under different optimality
criterions. We prove complexity upper bounds that are similar to standard
results established for convex functions and much better that state-of-the-art
convergence rates of non-convex functions. Overall, this paper suggests that
\textit{quasar-convexity} allows efficient optimization procedures, and we are
looking forward to seeing more problems that demonstrate similar properties in
practice.
Related papers
- A Novel Unified Parametric Assumption for Nonconvex Optimization [53.943470475510196]
Non optimization is central to machine learning, but the general framework non convexity enables weak convergence guarantees too pessimistic compared to the other hand.
We introduce a novel unified assumption in non convex algorithms.
arXiv Detail & Related papers (2025-02-17T21:25:31Z) - Extended convexity and smoothness and their applications in deep learning [5.281849820329249]
This paper introduces an optimization framework aimed at providing a theoretical foundation for a class of composite optimization problems, particularly those in deep learning.
We analyze the smoothness of Lipschitz's concepts of Lipschitz's descent and descent methods for objective functions that are $mathcalH(Phi)$-smoothness.
arXiv Detail & Related papers (2024-10-08T08:40:07Z) - Linearization Algorithms for Fully Composite Optimization [61.20539085730636]
This paper studies first-order algorithms for solving fully composite optimization problems convex compact sets.
We leverage the structure of the objective by handling differentiable and non-differentiable separately, linearizing only the smooth parts.
arXiv Detail & Related papers (2023-02-24T18:41:48Z) - High-Probability Bounds for Stochastic Optimization and Variational
Inequalities: the Case of Unbounded Variance [59.211456992422136]
We propose algorithms with high-probability convergence results under less restrictive assumptions.
These results justify the usage of the considered methods for solving problems that do not fit standard functional classes in optimization.
arXiv Detail & Related papers (2023-02-02T10:37:23Z) - Faster Accelerated First-order Methods for Convex Optimization with Strongly Convex Function Constraints [3.1044138971639734]
We introduce faster accelerated primal-dual algorithms for minimizing a convex function subject to strongly convex function constraints.
We show the superior performance our methods in sparsity-inducing optimization, notably Google's personalized PageRank problem.
arXiv Detail & Related papers (2022-12-21T16:04:53Z) - Recent Theoretical Advances in Non-Convex Optimization [56.88981258425256]
Motivated by recent increased interest in analysis of optimization algorithms for non- optimization in deep networks and other problems in data, we give an overview of recent results of theoretical optimization algorithms for non- optimization.
arXiv Detail & Related papers (2020-12-11T08:28:51Z) - Efficient Methods for Structured Nonconvex-Nonconcave Min-Max
Optimization [98.0595480384208]
We propose a generalization extraient spaces which converges to a stationary point.
The algorithm applies not only to general $p$-normed spaces, but also to general $p$-dimensional vector spaces.
arXiv Detail & Related papers (2020-10-31T21:35:42Z) - Fast Objective & Duality Gap Convergence for Non-Convex Strongly-Concave
Min-Max Problems with PL Condition [52.08417569774822]
This paper focuses on methods for solving smooth non-concave min-max problems, which have received increasing attention due to deep learning (e.g., deep AUC)
arXiv Detail & Related papers (2020-06-12T00:32:21Z) - Do optimization methods in deep learning applications matter? [0.0]
The paper presents arguments on which optimization functions to use and further, which functions would benefit from parallelization efforts.
Our experiments compare off-the-shelf optimization functions(CG, SGD, LM and L-BFGS) in standard CIFAR, MNIST, CartPole and FlappyBird experiments.
arXiv Detail & Related papers (2020-02-28T10:36:40Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.