The connections between Lyapunov functions for some optimization
algorithms and differential equations
- URL: http://arxiv.org/abs/2009.00673v2
- Date: Mon, 11 Jan 2021 11:18:40 GMT
- Title: The connections between Lyapunov functions for some optimization
algorithms and differential equations
- Authors: J.M. Sanz-Serna and Konstantinos C. Zygalakis
- Abstract summary: We derive analytically a (discrete) Lyapunov function for a family of Nesterov optimization methods.
We show that the majority of typical discretizations of the family of ODEs, such as the Heavy ball method, do not possess Lyapunov functions with properties similar to those of the Lyapunov function constructed here.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this manuscript, we study the properties of a family of second-order
differential equations with damping, its discretizations and their connections
with accelerated optimization algorithms for $m$-strongly convex and $L$-smooth
functions. In particular, using the Linear Matrix Inequality LMI framework
developed by \emph{Fazlyab et. al. $(2018)$}, we derive analytically a
(discrete) Lyapunov function for a two-parameter family of Nesterov
optimization methods, which allows for the complete characterization of their
convergence rate. In the appropriate limit, this family of methods may be seen
as a discretization of a family of second-order ordinary differential equations
for which we construct(continuous) Lyapunov functions by means of the LMI
framework. The continuous Lyapunov functions may alternatively, be obtained by
studying the limiting behaviour of their discrete counterparts. Finally, we
show that the majority of typical discretizations of the family of ODEs, such
as the Heavy ball method, do not possess Lyapunov functions with properties
similar to those of the Lyapunov function constructed here for the Nesterov
method.
Related papers
- Quantum algorithm for linear non-unitary dynamics with near-optimal
dependence on all parameters [5.471398994781521]
We introduce a family of identities that express general linear non-unitary evolution operators as a linear combination of unitary evolution operators.
This formulation can exponentially enhance the accuracy of the recently introduced linear combination of Hamiltonian simulation (LCHS) method.
arXiv Detail & Related papers (2023-12-06T21:30:26Z) - GloptiNets: Scalable Non-Convex Optimization with Certificates [61.50835040805378]
We present a novel approach to non-cube optimization with certificates, which handles smooth functions on the hypercube or on the torus.
By exploiting the regularity of the target function intrinsic in the decay of its spectrum, we allow at the same time to obtain precise certificates and leverage the advanced and powerful neural networks.
arXiv Detail & Related papers (2023-06-26T09:42:59Z) - On the connections between optimization algorithms, Lyapunov functions, and differential equations: theory and insights [0.0]
We revisit the framework introduced by Fazylab et al. to construct Lyapunov functions for optimization algorithms in discrete and continuous time.
For smooth, strongly convex objective functions, we relax the requirements necessary for such a construction.
We prove convergence rates that improve on those available in the literature.
arXiv Detail & Related papers (2023-05-15T14:03:16Z) - Understanding Accelerated Gradient Methods: Lyapunov Analyses and
Hamiltonian Assisted Interpretations [1.0152838128195465]
We formulate two classes of first-order algorithms more general than previously studied for minimizing smooth and strongly convex functions.
We establish sufficient conditions, via new discrete Lyapunov analyses, for achieving accelerated convergence rates which match Nesterov's methods in the strongly and general convex settings.
We propose a novel class of discrete algorithms, called the Hamiltonian assisted gradient method, directly based on a Hamiltonian function and several interpretable operations.
arXiv Detail & Related papers (2023-04-20T03:03:30Z) - An inexact LPA for DC composite optimization and application to matrix completions with outliers [5.746154410100363]
This paper concerns a class of composite optimization problems.
By leveraging the composite structure, we provide a condition for the potential function to have the KL property of $1/2$ at the iterate sequence.
arXiv Detail & Related papers (2023-03-29T16:15:34Z) - Linearization Algorithms for Fully Composite Optimization [61.20539085730636]
This paper studies first-order algorithms for solving fully composite optimization problems convex compact sets.
We leverage the structure of the objective by handling differentiable and non-differentiable separately, linearizing only the smooth parts.
arXiv Detail & Related papers (2023-02-24T18:41:48Z) - Reinforcement Learning from Partial Observation: Linear Function Approximation with Provable Sample Efficiency [111.83670279016599]
We study reinforcement learning for partially observed decision processes (POMDPs) with infinite observation and state spaces.
We make the first attempt at partial observability and function approximation for a class of POMDPs with a linear structure.
arXiv Detail & Related papers (2022-04-20T21:15:38Z) - A Discrete Variational Derivation of Accelerated Methods in Optimization [68.8204255655161]
We introduce variational which allow us to derive different methods for optimization.
We derive two families of optimization methods in one-to-one correspondence.
The preservation of symplecticity of autonomous systems occurs here solely on the fibers.
arXiv Detail & Related papers (2021-06-04T20:21:53Z) - Automated and Sound Synthesis of Lyapunov Functions with SMT Solvers [70.70479436076238]
We synthesise Lyapunov functions for linear, non-linear (polynomial) and parametric models.
We exploit an inductive framework to synthesise Lyapunov functions, starting from parametric templates.
arXiv Detail & Related papers (2020-07-21T14:45:23Z) - Optimal Bounds between $f$-Divergences and Integral Probability Metrics [8.401473551081748]
Families of $f$-divergences and Integral Probability Metrics are widely used to quantify similarity between probability distributions.
We systematically study the relationship between these two families from the perspective of convex duality.
We obtain new bounds while also recovering in a unified manner well-known results, such as Hoeffding's lemma.
arXiv Detail & Related papers (2020-06-10T17:39:11Z) - Formal Synthesis of Lyapunov Neural Networks [61.79595926825511]
We propose an automatic and formally sound method for synthesising Lyapunov functions.
We employ a counterexample-guided approach where a numerical learner and a symbolic verifier interact to construct provably correct Lyapunov neural networks.
Our method synthesises Lyapunov functions faster and over wider spatial domains than the alternatives, yet providing stronger or equal guarantees.
arXiv Detail & Related papers (2020-03-19T17:21:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.