Complementary Composite Minimization, Small Gradients in General Norms,
and Applications to Regression Problems
- URL: http://arxiv.org/abs/2101.11041v1
- Date: Tue, 26 Jan 2021 19:21:28 GMT
- Title: Complementary Composite Minimization, Small Gradients in General Norms,
and Applications to Regression Problems
- Authors: Jelena Diakonikolas and Crist\'obal Guzm\'an
- Abstract summary: Composite minimization is a powerful framework in large-scale convex optimization.
We introduce a new algorithmic framework for complementary composite minimization.
We prove that the algorithms resulting from our framework are near-optimal in most of the standard optimization settings.
- Score: 14.759688428864157
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Composite minimization is a powerful framework in large-scale convex
optimization, based on decoupling of the objective function into terms with
structurally different properties and allowing for more flexible algorithmic
design. In this work, we introduce a new algorithmic framework for
complementary composite minimization, where the objective function decouples
into a (weakly) smooth and a uniformly convex term. This particular form of
decoupling is pervasive in statistics and machine learning, due to its link to
regularization.
The main contributions of our work are summarized as follows. First, we
introduce the problem of complementary composite minimization in general normed
spaces; second, we provide a unified accelerated algorithmic framework to
address broad classes of complementary composite minimization problems; and
third, we prove that the algorithms resulting from our framework are
near-optimal in most of the standard optimization settings. Additionally, we
show that our algorithmic framework can be used to address the problem of
making the gradients small in general normed spaces. As a concrete example, we
obtain a nearly-optimal method for the standard $\ell_1$ setup (small gradients
in the $\ell_\infty$ norm), essentially matching the bound of Nesterov (2012)
that was previously known only for the Euclidean setup. Finally, we show that
our composite methods are broadly applicable to a number of regression
problems, leading to complexity bounds that are either new or match the best
existing ones.
Related papers
- Minimizing Quasi-Self-Concordant Functions by Gradient Regularization of
Newton Method [4.62316736194615]
We study the composite convex optimization problems with a Quasi-Self-Concordant smooth component.
This problem class naturally interpolates between classic Self-Concordant functions and functions with Lipschitz continuous Hessian.
We show that for minimizing Quasi-Self-Concordant functions we can use instead the basic Newton Method with Gradient Regularization.
arXiv Detail & Related papers (2023-08-28T17:43:04Z) - Linearization Algorithms for Fully Composite Optimization [61.20539085730636]
This paper studies first-order algorithms for solving fully composite optimization problems convex compact sets.
We leverage the structure of the objective by handling differentiable and non-differentiable separately, linearizing only the smooth parts.
arXiv Detail & Related papers (2023-02-24T18:41:48Z) - Optimal Algorithms for Stochastic Complementary Composite Minimization [55.26935605535377]
Inspired by regularization techniques in statistics and machine learning, we study complementary composite minimization.
We provide novel excess risk bounds, both in expectation and with high probability.
Our algorithms are nearly optimal, which we prove via novel lower complexity bounds for this class of problems.
arXiv Detail & Related papers (2022-11-03T12:40:24Z) - A Constrained Optimization Approach to Bilevel Optimization with
Multiple Inner Minima [49.320758794766185]
We propose a new approach, which convert the bilevel problem to an equivalent constrained optimization, and then the primal-dual algorithm can be used to solve the problem.
Such an approach enjoys a few advantages including (a) addresses the multiple inner minima challenge; (b) fully first-order efficiency without Jacobian computations.
arXiv Detail & Related papers (2022-03-01T18:20:01Z) - SCORE: Approximating Curvature Information under Self-Concordant
Regularization [0.0]
We propose a generalized Gauss-Newton with Self-Concordant Regularization (GGN-SCORE) algorithm that updates the minimization speed each time it receives a new input.
The proposed algorithm exploits the structure of the second-order information in the Hessian matrix, thereby reducing computational overhead.
arXiv Detail & Related papers (2021-12-14T13:03:04Z) - On Constraints in First-Order Optimization: A View from Non-Smooth
Dynamical Systems [99.59934203759754]
We introduce a class of first-order methods for smooth constrained optimization.
Two distinctive features of our approach are that projections or optimizations over the entire feasible set are avoided.
The resulting algorithmic procedure is simple to implement even when constraints are nonlinear.
arXiv Detail & Related papers (2021-07-17T11:45:13Z) - Efficient Methods for Structured Nonconvex-Nonconcave Min-Max
Optimization [98.0595480384208]
We propose a generalization extraient spaces which converges to a stationary point.
The algorithm applies not only to general $p$-normed spaces, but also to general $p$-dimensional vector spaces.
arXiv Detail & Related papers (2020-10-31T21:35:42Z) - Conditional gradient methods for stochastically constrained convex
minimization [54.53786593679331]
We propose two novel conditional gradient-based methods for solving structured convex optimization problems.
The most important feature of our framework is that only a subset of the constraints is processed at each iteration.
Our algorithms rely on variance reduction and smoothing used in conjunction with conditional gradient steps, and are accompanied by rigorous convergence guarantees.
arXiv Detail & Related papers (2020-07-07T21:26:35Z) - A Class of Linear Programs Solvable by Coordinate-Wise Minimization [0.0]
We present a class of linear programs that coordinate-wise minimization solves exactly.
We show that dual LP relaxations of several well-known optimization problems are in this class.
Our results are theoretically non-trivial and can lead to new large-scale optimization algorithms in the future.
arXiv Detail & Related papers (2020-01-28T17:14:47Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.