An Adaptive Stochastic Sequential Quadratic Programming with
Differentiable Exact Augmented Lagrangians
- URL: http://arxiv.org/abs/2102.05320v1
- Date: Wed, 10 Feb 2021 08:40:55 GMT
- Title: An Adaptive Stochastic Sequential Quadratic Programming with
Differentiable Exact Augmented Lagrangians
- Authors: Sen Na, Mihai Anitescu, Mladen Kolar
- Abstract summary: We consider the problem of solving nonlinear optimization programs with objective and deterministic equality.
We propose a sequential quadratic programming (SQP) that uses a differentiable exact augmented Lagrangian as the merit function.
The proposed algorithm is the first SQP that allows a line search procedure and the first line search procedure.
- Score: 17.9230793188835
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: We consider the problem of solving nonlinear optimization programs with
stochastic objective and deterministic equality constraints. We assume for the
objective that the function evaluation, the gradient, and the Hessian are
inaccessible, while one can compute their stochastic estimates by, for example,
subsampling. We propose a stochastic algorithm based on sequential quadratic
programming (SQP) that uses a differentiable exact augmented Lagrangian as the
merit function. To motivate our algorithm, we revisit an old SQP method
\citep{Lucidi1990Recursive} developed for deterministic programs. We simplify
that method and derive an adaptive SQP, which serves as the skeleton of our
stochastic algorithm. Based on the derived algorithm, we then propose a
non-adaptive SQP for optimizing stochastic objectives, where the gradient and
the Hessian are replaced by stochastic estimates but the stepsize is
deterministic and prespecified. Finally, we incorporate a recent stochastic
line search procedure \citep{Paquette2020Stochastic} into our non-adaptive
stochastic SQP to arrive at an adaptive stochastic SQP. To our knowledge, the
proposed algorithm is the first stochastic SQP that allows a line search
procedure and the first stochastic line search procedure that allows the
constraints. The global convergence for all proposed SQP methods is
established, while numerical experiments on nonlinear problems in the CUTEst
test set demonstrate the superiority of the proposed algorithm.
Related papers
- Trust-Region Sequential Quadratic Programming for Stochastic Optimization with Random Models [57.52124921268249]
We propose a Trust Sequential Quadratic Programming method to find both first and second-order stationary points.
To converge to first-order stationary points, our method computes a gradient step in each iteration defined by minimizing a approximation of the objective subject.
To converge to second-order stationary points, our method additionally computes an eigen step to explore the negative curvature the reduced Hessian matrix.
arXiv Detail & Related papers (2024-09-24T04:39:47Z) - A Sequential Quadratic Programming Method with High Probability Complexity Bounds for Nonlinear Equality Constrained Stochastic Optimization [2.3814052021083354]
It is assumed that constraint function values and derivatives are available, but only programming approximations of the objective function and its associated derivatives can be computed.
A high-probability bound on the iteration complexity of the algorithm to approximate first-order stationarity is derived.
arXiv Detail & Related papers (2023-01-01T21:46:50Z) - Fully Stochastic Trust-Region Sequential Quadratic Programming for
Equality-Constrained Optimization Problems [62.83783246648714]
We propose a sequential quadratic programming algorithm (TR-StoSQP) to solve nonlinear optimization problems with objectives and deterministic equality constraints.
The algorithm adaptively selects the trust-region radius and, compared to the existing line-search StoSQP schemes, allows us to utilize indefinite Hessian matrices.
arXiv Detail & Related papers (2022-11-29T05:52:17Z) - Maximum-Likelihood Inverse Reinforcement Learning with Finite-Time
Guarantees [56.848265937921354]
Inverse reinforcement learning (IRL) aims to recover the reward function and the associated optimal policy.
Many algorithms for IRL have an inherently nested structure.
We develop a novel single-loop algorithm for IRL that does not compromise reward estimation accuracy.
arXiv Detail & Related papers (2022-10-04T17:13:45Z) - Statistical Inference of Constrained Stochastic Optimization via Sketched Sequential Quadratic Programming [53.63469275932989]
We consider online statistical inference of constrained nonlinear optimization problems.
We apply the Sequential Quadratic Programming (StoSQP) method to solve these problems.
arXiv Detail & Related papers (2022-05-27T00:34:03Z) - Inequality Constrained Stochastic Nonlinear Optimization via Active-Set
Sequential Quadratic Programming [17.9230793188835]
We study nonlinear optimization problems with objective and deterministic equality and inequality constraints.
We propose an active-set sequentialAdaptive programming algorithm, using a differentiable exact augmented Lagrangian as the merit function.
The algorithm adaptively selects the parameters of augmented Lagrangian and performs line search to decide the stepsize.
arXiv Detail & Related papers (2021-09-23T17:12:17Z) - Zeroth and First Order Stochastic Frank-Wolfe Algorithms for Constrained
Optimization [13.170519806372075]
Problems of convex optimization with two sets of constraints arise frequently in the context of semidefinite programming.
Since projection onto the first set of constraints is difficult, it becomes necessary to explore projection-free algorithms.
The efficacy of the proposed algorithms is tested on relevant applications of sparse matrix estimation, clustering via semidefinite relaxation, and uniform sparsest cut problem.
arXiv Detail & Related papers (2021-07-14T08:01:30Z) - High Probability Complexity Bounds for Non-Smooth Stochastic Optimization with Heavy-Tailed Noise [51.31435087414348]
It is essential to theoretically guarantee that algorithms provide small objective residual with high probability.
Existing methods for non-smooth convex optimization have complexity bounds with dependence on confidence level.
We propose novel stepsize rules for two methods with gradient clipping.
arXiv Detail & Related papers (2021-06-10T17:54:21Z) - Sequential Quadratic Optimization for Nonlinear Equality Constrained
Stochastic Optimization [10.017195276758454]
It is assumed in this setting that it is intractable to compute objective function and derivative values explicitly.
An algorithm is proposed for the deterministic setting that is modeled after a state-of-the-art line-search SQP algorithm.
The results of numerical experiments demonstrate the practical performance of our proposed techniques.
arXiv Detail & Related papers (2020-07-20T23:04:26Z) - Convergence of adaptive algorithms for weakly convex constrained
optimization [59.36386973876765]
We prove the $mathcaltilde O(t-1/4)$ rate of convergence for the norm of the gradient of Moreau envelope.
Our analysis works with mini-batch size of $1$, constant first and second order moment parameters, and possibly smooth optimization domains.
arXiv Detail & Related papers (2020-06-11T17:43:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.