Stochastic Variable Metric Proximal Gradient with variance reduction for
non-convex composite optimization
- URL: http://arxiv.org/abs/2301.00631v1
- Date: Mon, 2 Jan 2023 12:49:48 GMT
- Title: Stochastic Variable Metric Proximal Gradient with variance reduction for
non-convex composite optimization
- Authors: Gersende Fort (IMT), Eric Moulines (CMAP)
- Abstract summary: 3P-SP-IDER is a novel algorithm designed to solve finite sum non-backward logistic equations.
We show that 3P-SP-IDER extends some preconditioned and some Incremental Maximization algorithms to the case forward operator can not be computed in closed form.
We also discuss the role of some design parameters of 3P-SP-IDER in some application to inference in a regression model with random effects.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper introduces a novel algorithm, the Perturbed Proximal
Preconditioned SPIDER algorithm (3P-SPIDER), designed to solve finite sum
non-convex composite optimization. It is a stochastic Variable Metric
Forward-Backward algorithm, which allows approximate preconditioned forward
operator and uses a variable metric proximity operator as the backward
operator; it also proposes a mini-batch strategy with variance reduction to
address the finite sum setting. We show that 3P-SPIDER extends some Stochastic
preconditioned Gradient Descent-based algorithms and some Incremental
Expectation Maximization algorithms to composite optimization and to the case
the forward operator can not be computed in closed form. We also provide an
explicit control of convergence in expectation of 3P-SPIDER, and study its
complexity in order to satisfy the epsilon-approximate stationary condition.
Our results are the first to combine the composite non-convex optimization
setting, a variance reduction technique to tackle the finite sum setting by
using a minibatch strategy and, to allow deterministic or random approximations
of the preconditioned forward operator. Finally, through an application to
inference in a logistic regression model with random effects, we numerically
compare 3P-SPIDER to other stochastic forward-backward algorithms and discuss
the role of some design parameters of 3P-SPIDER.
Related papers
- Variable Substitution and Bilinear Programming for Aligning Partially Overlapping Point Sets [48.1015832267945]
This research presents a method to meet requirements through the minimization objective function of the RPM algorithm.
A branch-and-bound (BnB) algorithm is devised, which solely branches over the parameters, thereby boosting convergence rate.
Empirical evaluations demonstrate better robustness of the proposed methodology against non-rigid deformation, positional noise, and outliers, when compared with prevailing state-of-the-art transformations.
arXiv Detail & Related papers (2024-05-14T13:28:57Z) - Reducing measurement costs by recycling the Hessian in adaptive
variational quantum algorithms [0.0]
We propose an improved quasi-Newton optimization protocol specifically tailored to adaptive VQAs.
We implement a quasi-Newton algorithm where an approximation to the inverse Hessian matrix is continuously built and grown across the iterations of an adaptive VQA.
arXiv Detail & Related papers (2024-01-10T14:08:04Z) - Stochastic Optimization for Non-convex Problem with Inexact Hessian
Matrix, Gradient, and Function [99.31457740916815]
Trust-region (TR) and adaptive regularization using cubics have proven to have some very appealing theoretical properties.
We show that TR and ARC methods can simultaneously provide inexact computations of the Hessian, gradient, and function values.
arXiv Detail & Related papers (2023-10-18T10:29:58Z) - Momentum Accelerates the Convergence of Stochastic AUPRC Maximization [80.8226518642952]
We study optimization of areas under precision-recall curves (AUPRC), which is widely used for imbalanced tasks.
We develop novel momentum methods with a better iteration of $O (1/epsilon4)$ for finding an $epsilon$stationary solution.
We also design a novel family of adaptive methods with the same complexity of $O (1/epsilon4)$, which enjoy faster convergence in practice.
arXiv Detail & Related papers (2021-07-02T16:21:52Z) - The Perturbed Prox-Preconditioned SPIDER algorithm for EM-based large
scale learning [0.0]
Incremental Expectation Maximization (EM) algorithms were introduced to design EM for the large scale learning framework.
We propose a novel algorithm named Perturbed Prox-Preconditioned SPIDER, which builds on the Perturbed Integral EstimatoR EM (SPIDER-EM) algorithm.
The 3P-SPIDER algorithm addresses many intractabilities of the E-step of EM; it also deals with non-smooth regularization and convex constraint set.
arXiv Detail & Related papers (2021-05-25T07:54:58Z) - On Stochastic Moving-Average Estimators for Non-Convex Optimization [105.22760323075008]
In this paper, we demonstrate the power of a widely used estimator based on moving average (SEMA) problems.
For all these-the-art results, we also present the results for all these-the-art problems.
arXiv Detail & Related papers (2021-04-30T08:50:24Z) - Geom-SPIDER-EM: Faster Variance Reduced Stochastic Expectation
Maximization for Nonconvex Finite-Sum Optimization [21.81837334970773]
We propose an extension of the Path-Integrated Differential Estima to the Expectation Maximization (EM) algorithm.
We show it supports the same state art bounds as SPIDER-EM-IDER; and results provide for a rate for our findings.
arXiv Detail & Related papers (2020-11-24T21:20:53Z) - Sparse Representations of Positive Functions via First and Second-Order
Pseudo-Mirror Descent [15.340540198612823]
We consider expected risk problems when the range of the estimator is required to be nonnegative.
We develop first and second-order variants of approximation mirror descent employing emphpseudo-gradients.
Experiments demonstrate favorable performance on ingeneous Process intensity estimation in practice.
arXiv Detail & Related papers (2020-11-13T21:54:28Z) - Adaptive pruning-based optimization of parameterized quantum circuits [62.997667081978825]
Variisy hybrid quantum-classical algorithms are powerful tools to maximize the use of Noisy Intermediate Scale Quantum devices.
We propose a strategy for such ansatze used in variational quantum algorithms, which we call "Efficient Circuit Training" (PECT)
Instead of optimizing all of the ansatz parameters at once, PECT launches a sequence of variational algorithms.
arXiv Detail & Related papers (2020-10-01T18:14:11Z) - Convergence of adaptive algorithms for weakly convex constrained
optimization [59.36386973876765]
We prove the $mathcaltilde O(t-1/4)$ rate of convergence for the norm of the gradient of Moreau envelope.
Our analysis works with mini-batch size of $1$, constant first and second order moment parameters, and possibly smooth optimization domains.
arXiv Detail & Related papers (2020-06-11T17:43:19Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.