Stochastic Multi-level Composition Optimization Algorithms with
Level-Independent Convergence Rates
- URL: http://arxiv.org/abs/2008.10526v5
- Date: Mon, 14 Feb 2022 05:55:45 GMT
- Title: Stochastic Multi-level Composition Optimization Algorithms with
Level-Independent Convergence Rates
- Authors: Krishnakumar Balasubramanian, Saeed Ghadimi, Anthony Nguyen
- Abstract summary: We study smooth multi-level composition optimization problems, where the objective function is a nested composition of $T$ functions.
We show that the first algorithm, which is a generalization of citeGhaRuswan20 to the $T$ level case, can achieve a sample complexity of $mathcalO (1/epsilon$6)
This is the first time that such an online algorithm designed for the (un) multi-level setting, obtains the same sample complexity under standard assumptions.
- Score: 12.783783498844022
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper, we study smooth stochastic multi-level composition
optimization problems, where the objective function is a nested composition of
$T$ functions. We assume access to noisy evaluations of the functions and their
gradients, through a stochastic first-order oracle. For solving this class of
problems, we propose two algorithms using moving-average stochastic estimates,
and analyze their convergence to an $\epsilon$-stationary point of the problem.
We show that the first algorithm, which is a generalization of
\cite{GhaRuswan20} to the $T$ level case, can achieve a sample complexity of
$\mathcal{O}(1/\epsilon^6)$ by using mini-batches of samples in each iteration.
By modifying this algorithm using linearized stochastic estimates of the
function values, we improve the sample complexity to
$\mathcal{O}(1/\epsilon^4)$. {\color{black}This modification not only removes
the requirement of having a mini-batch of samples in each iteration, but also
makes the algorithm parameter-free and easy to implement}. To the best of our
knowledge, this is the first time that such an online algorithm designed for
the (un)constrained multi-level setting, obtains the same sample complexity of
the smooth single-level setting, under standard assumptions (unbiasedness and
boundedness of the second moments) on the stochastic first-order oracle.
Related papers
- Fully Zeroth-Order Bilevel Programming via Gaussian Smoothing [7.143879014059895]
We study and analyze zeroth-order approximation algorithms for solving bilvel problems.
To the best of our knowledge, this is the first time that sample bounds are established for a fully zeroth-order bilevel optimization algorithm.
arXiv Detail & Related papers (2024-03-29T21:12:25Z) - An Oblivious Stochastic Composite Optimization Algorithm for Eigenvalue
Optimization Problems [76.2042837251496]
We introduce two oblivious mirror descent algorithms based on a complementary composite setting.
Remarkably, both algorithms work without prior knowledge of the Lipschitz constant or smoothness of the objective function.
We show how to extend our framework to scale and demonstrate the efficiency and robustness of our methods on large scale semidefinite programs.
arXiv Detail & Related papers (2023-06-30T08:34:29Z) - Multi-block-Single-probe Variance Reduced Estimator for Coupled
Compositional Optimization [49.58290066287418]
We propose a novel method named Multi-block-probe Variance Reduced (MSVR) to alleviate the complexity of compositional problems.
Our results improve upon prior ones in several aspects, including the order of sample complexities and dependence on strongity.
arXiv Detail & Related papers (2022-07-18T12:03:26Z) - A Projection-free Algorithm for Constrained Stochastic Multi-level
Composition Optimization [12.096252285460814]
We propose a projection-free conditional gradient-type algorithm for composition optimization.
We show that the number of oracles and the linear-minimization oracle required by the proposed algorithm, are of order $mathcalO_T(epsilon-2)$ and $mathcalO_T(epsilon-3)$ respectively.
arXiv Detail & Related papers (2022-02-09T06:05:38Z) - Bilevel Optimization with a Lower-level Contraction: Optimal Sample
Complexity without Warm-start [39.13249645102326]
We study bilevel problems in which the upper-level problem consists in the iterations of a smooth objective function and the lower-level problem is to find the fixed point of a smooth map.
Several recent works have proposed algorithms which warm-start the lower-level problem.
We show that without warm-start, it is still possible to achieve order-wise optimal sample complexity.
arXiv Detail & Related papers (2022-02-07T18:35:46Z) - Randomized Stochastic Variance-Reduced Methods for Stochastic Bilevel
Optimization [62.87181271021217]
We consider non-SBO problems that have many applications in machine learning.
This paper proposes fast randomized algorithms for non-SBO problems.
arXiv Detail & Related papers (2021-05-05T18:28:42Z) - A Momentum-Assisted Single-Timescale Stochastic Approximation Algorithm
for Bilevel Optimization [112.59170319105971]
We propose a new algorithm -- the Momentum- Single-timescale Approximation (MSTSA) -- for tackling problems.
MSTSA allows us to control the error in iterations due to inaccurate solution to the lower level subproblem.
arXiv Detail & Related papers (2021-02-15T07:10:33Z) - Single-Timescale Stochastic Nonconvex-Concave Optimization for Smooth
Nonlinear TD Learning [145.54544979467872]
We propose two single-timescale single-loop algorithms that require only one data point each step.
Our results are expressed in a form of simultaneous primal and dual side convergence.
arXiv Detail & Related papers (2020-08-23T20:36:49Z) - A Two-Timescale Framework for Bilevel Optimization: Complexity Analysis
and Application to Actor-Critic [142.1492359556374]
Bilevel optimization is a class of problems which exhibit a two-level structure.
We propose a two-timescale approximation (TTSA) algorithm for tackling such a bilevel problem.
We show that a two-timescale natural actor-critic policy optimization algorithm can be viewed as a special case of our TTSA framework.
arXiv Detail & Related papers (2020-07-10T05:20:02Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.