Blockwise Stochastic Variance-Reduced Methods with Parallel Speedup for
Multi-Block Bilevel Optimization
- URL: http://arxiv.org/abs/2305.18730v2
- Date: Fri, 2 Jun 2023 04:16:51 GMT
- Title: Blockwise Stochastic Variance-Reduced Methods with Parallel Speedup for
Multi-Block Bilevel Optimization
- Authors: Quanqi Hu, Zi-Hao Qiu, Zhishuai Guo, Lijun Zhang, Tianbao Yang
- Abstract summary: Non-stationary multi-block bilevel optimization problems involve $mgg 1$ lower level problems and have important applications in machine learning.
We aim to achieve three properties for our algorithm: a) matching the state-of-the-art complexity of standard BO problems with a single block; (b) achieving parallel speedup by sampling $I$ samples for each sampled block per-iteration; and (c) avoiding the computation of the inverse of a high-dimensional Hessian matrix estimator.
- Score: 43.74656748515853
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: In this paper, we consider non-convex multi-block bilevel optimization (MBBO)
problems, which involve $m\gg 1$ lower level problems and have important
applications in machine learning. Designing a stochastic gradient and
controlling its variance is more intricate due to the hierarchical sampling of
blocks and data and the unique challenge of estimating hyper-gradient. We aim
to achieve three nice properties for our algorithm: (a) matching the
state-of-the-art complexity of standard BO problems with a single block; (b)
achieving parallel speedup by sampling $I$ blocks and sampling $B$ samples for
each sampled block per-iteration; (c) avoiding the computation of the inverse
of a high-dimensional Hessian matrix estimator. However, it is non-trivial to
achieve all of these by observing that existing works only achieve one or two
of these properties. To address the involved challenges for achieving (a, b,
c), we propose two stochastic algorithms by using advanced blockwise
variance-reduction techniques for tracking the Hessian matrices (for
low-dimensional problems) or the Hessian-vector products (for high-dimensional
problems), and prove an iteration complexity of
$O(\frac{m\epsilon^{-3}\mathbb{I}(I<m)}{I\sqrt{I}} +
\frac{m\epsilon^{-3}}{I\sqrt{B}})$ for finding an $\epsilon$-stationary point
under appropriate conditions. We also conduct experiments to verify the
effectiveness of the proposed algorithms comparing with existing MBBO
algorithms.
Related papers
- Stochastic First-Order Methods with Non-smooth and Non-Euclidean Proximal Terms for Nonconvex High-Dimensional Stochastic Optimization [2.0657831823662574]
When the non problem is by which the non problem is by whichity, the sample of first-order methods may depend linearly on the problem dimension, is for undesirable problems.
Our algorithms allow for the estimate of complexity using the distance of.
mathO (log d) / EuM4.
We prove that DISFOM can sharpen variance employing $mathO (log d) / EuM4.
arXiv Detail & Related papers (2024-06-27T18:38:42Z) - Multi-block-Single-probe Variance Reduced Estimator for Coupled
Compositional Optimization [49.58290066287418]
We propose a novel method named Multi-block-probe Variance Reduced (MSVR) to alleviate the complexity of compositional problems.
Our results improve upon prior ones in several aspects, including the order of sample complexities and dependence on strongity.
arXiv Detail & Related papers (2022-07-18T12:03:26Z) - Quantum Goemans-Williamson Algorithm with the Hadamard Test and
Approximate Amplitude Constraints [62.72309460291971]
We introduce a variational quantum algorithm for Goemans-Williamson algorithm that uses only $n+1$ qubits.
Efficient optimization is achieved by encoding the objective matrix as a properly parameterized unitary conditioned on an auxilary qubit.
We demonstrate the effectiveness of our protocol by devising an efficient quantum implementation of the Goemans-Williamson algorithm for various NP-hard problems.
arXiv Detail & Related papers (2022-06-30T03:15:23Z) - Multi-block Min-max Bilevel Optimization with Applications in Multi-task
Deep AUC Maximization [36.743632418094]
We present a single-loop algorithm to tackle a multiblock min-max bilevel optimization problem.
We show that our algorithm can be used to tackle problems with hundreds of tasks.
arXiv Detail & Related papers (2022-06-01T06:42:36Z) - Randomized Stochastic Variance-Reduced Methods for Stochastic Bilevel
Optimization [62.87181271021217]
We consider non-SBO problems that have many applications in machine learning.
This paper proposes fast randomized algorithms for non-SBO problems.
arXiv Detail & Related papers (2021-05-05T18:28:42Z) - Hybrid Trilinear and Bilinear Programming for Aligning Partially
Overlapping Point Sets [85.71360365315128]
In many applications, we need algorithms which can align partially overlapping point sets are invariant to the corresponding corresponding RPM algorithm.
We first show that the objective is a cubic bound function. We then utilize the convex envelopes of trilinear and bilinear monomial transformations to derive its lower bound.
We next develop a branch-and-bound (BnB) algorithm which only branches over the transformation variables and runs efficiently.
arXiv Detail & Related papers (2021-01-19T04:24:23Z) - Block majorization-minimization with diminishing radius for constrained
nonconvex optimization [9.907540661545328]
Block tensor regularization-minimization (BMM) is a simple iterative algorithm for non constrained optimization that minimizes major surrogates in each block.
We show that BMM can produce a gradient $O(epsilon-2(logepsilon-1)2)$ when convex surrogates are used.
arXiv Detail & Related papers (2020-12-07T07:53:09Z) - Single-Timescale Stochastic Nonconvex-Concave Optimization for Smooth
Nonlinear TD Learning [145.54544979467872]
We propose two single-timescale single-loop algorithms that require only one data point each step.
Our results are expressed in a form of simultaneous primal and dual side convergence.
arXiv Detail & Related papers (2020-08-23T20:36:49Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.