A Contraction Theory Approach to Optimization Algorithms from
Acceleration Flows
- URL: http://arxiv.org/abs/2105.08832v1
- Date: Tue, 18 May 2021 21:11:37 GMT
- Title: A Contraction Theory Approach to Optimization Algorithms from
Acceleration Flows
- Authors: Pedro Cisneros-Velarde, Francesco Bullo
- Abstract summary: We use contraction theory to provide a principled methodology to design and discretize appropriate ODEs.
We propose a novel system of ODEs, namely the Accelerated-Contracting-Nesterov flow.
Remarkably, a simple explicit Euler discretization of this flow corresponds to the Nesterov acceleration method.
- Score: 1.90365714903665
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Much recent interest has focused on the design of optimization algorithms
from the discretization of an associated optimization flow, i.e., a system of
differential equations (ODEs) whose trajectories solve an associated
optimization problem. Such a design approach poses an important problem: how to
find a principled methodology to design and discretize appropriate ODEs. This
paper aims to provide a solution to this problem through the use of contraction
theory. We first introduce general mathematical results that explain how
contraction theory guarantees the stability of the implicit and explicit Euler
integration methods. Then, we propose a novel system of ODEs, namely the
Accelerated-Contracting-Nesterov flow, and use contraction theory to establish
it is an optimization flow with exponential convergence rate, from which the
linear convergence rate of its associated optimization algorithm is immediately
established. Remarkably, a simple explicit Euler discretization of this flow
corresponds to the Nesterov acceleration method. Finally, we present how our
approach leads to performance guarantees in the design of optimization
algorithms for time-varying optimization problems.
Related papers
- ODE-based Learning to Optimize [28.380622776436905]
We present a comprehensive framework integrating the inertial systems with Hessian-driven damping equation (ISHD)
We formulate a novel learning to optimize (L2O) problem aimed at minimizing the stopping time subject to the convergence and stability condition.
Empirical validation of our framework is conducted through extensive numerical experiments across a diverse set of optimization problems.
arXiv Detail & Related papers (2024-06-04T06:39:45Z) - Analyzing and Enhancing the Backward-Pass Convergence of Unrolled
Optimization [50.38518771642365]
The integration of constrained optimization models as components in deep networks has led to promising advances on many specialized learning tasks.
A central challenge in this setting is backpropagation through the solution of an optimization problem, which often lacks a closed form.
This paper provides theoretical insights into the backward pass of unrolled optimization, showing that it is equivalent to the solution of a linear system by a particular iterative method.
A system called Folded Optimization is proposed to construct more efficient backpropagation rules from unrolled solver implementations.
arXiv Detail & Related papers (2023-12-28T23:15:18Z) - Accelerating Cutting-Plane Algorithms via Reinforcement Learning
Surrogates [49.84541884653309]
A current standard approach to solving convex discrete optimization problems is the use of cutting-plane algorithms.
Despite the existence of a number of general-purpose cut-generating algorithms, large-scale discrete optimization problems continue to suffer from intractability.
We propose a method for accelerating cutting-plane algorithms via reinforcement learning.
arXiv Detail & Related papers (2023-07-17T20:11:56Z) - Fast Computation of Optimal Transport via Entropy-Regularized Extragradient Methods [75.34939761152587]
Efficient computation of the optimal transport distance between two distributions serves as an algorithm that empowers various applications.
This paper develops a scalable first-order optimization-based method that computes optimal transport to within $varepsilon$ additive accuracy.
arXiv Detail & Related papers (2023-01-30T15:46:39Z) - Backpropagation of Unrolled Solvers with Folded Optimization [55.04219793298687]
The integration of constrained optimization models as components in deep networks has led to promising advances on many specialized learning tasks.
One typical strategy is algorithm unrolling, which relies on automatic differentiation through the operations of an iterative solver.
This paper provides theoretical insights into the backward pass of unrolled optimization, leading to a system for generating efficiently solvable analytical models of backpropagation.
arXiv Detail & Related papers (2023-01-28T01:50:42Z) - A Derivation of Nesterov's Accelerated Gradient Algorithm from Optimal
Control Theory [0.0]
Nesterov's accelerated gradient algorithm is derived from first principles.
An Euler discretization of the resulting differential equation produces Nesterov's algorithm.
arXiv Detail & Related papers (2022-03-29T19:26:20Z) - Breaking the Convergence Barrier: Optimization via Fixed-Time Convergent
Flows [4.817429789586127]
We introduce a Poly-based optimization framework for achieving acceleration, based on the notion of fixed-time stability dynamical systems.
We validate the accelerated convergence properties of the proposed schemes on a range of numerical examples against the state-of-the-art optimization algorithms.
arXiv Detail & Related papers (2021-12-02T16:04:40Z) - Optimization on manifolds: A symplectic approach [127.54402681305629]
We propose a dissipative extension of Dirac's theory of constrained Hamiltonian systems as a general framework for solving optimization problems.
Our class of (accelerated) algorithms are not only simple and efficient but also applicable to a broad range of contexts.
arXiv Detail & Related papers (2021-07-23T13:43:34Z) - First-Order Optimization Inspired from Finite-Time Convergent Flows [26.931390502212825]
We propose an Euler discretization for first-order finite-time flows, and provide convergence guarantees, in the deterministic and the deterministic setting.
We then apply the proposed algorithms to academic examples, as well as deep neural networks training, where we empirically test their performances on the SVHN dataset.
Our results show that our schemes demonstrate faster convergences against standard optimization alternatives.
arXiv Detail & Related papers (2020-10-06T19:28:00Z) - IDEAL: Inexact DEcentralized Accelerated Augmented Lagrangian Method [64.15649345392822]
We introduce a framework for designing primal methods under the decentralized optimization setting where local functions are smooth and strongly convex.
Our approach consists of approximately solving a sequence of sub-problems induced by the accelerated augmented Lagrangian method.
When coupled with accelerated gradient descent, our framework yields a novel primal algorithm whose convergence rate is optimal and matched by recently derived lower bounds.
arXiv Detail & Related papers (2020-06-11T18:49:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.