Value-Function-based Sequential Minimization for Bi-level Optimization
- URL: http://arxiv.org/abs/2110.04974v2
- Date: Sat, 6 May 2023 12:31:38 GMT
- Title: Value-Function-based Sequential Minimization for Bi-level Optimization
- Authors: Risheng Liu, Xuan Liu, Shangzhi Zeng, Jin Zhang, Yixuan Zhang
- Abstract summary: gradient-based Bi-Level Optimization (BLO) methods have been widely applied to handle modern learning tasks.
There are almost no gradient-based methods able to solve BLO in challenging scenarios, such as BLO with functional constraints and pessimistic BLO.
We provide Bi-level Value-Function-based Sequential Minimization (BVFSM) to address the above issues.
- Score: 52.39882976848064
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Gradient-based Bi-Level Optimization (BLO) methods have been widely applied
to handle modern learning tasks. However, most existing strategies are
theoretically designed based on restrictive assumptions (e.g., convexity of the
lower-level sub-problem), and computationally not applicable for
high-dimensional tasks. Moreover, there are almost no gradient-based methods
able to solve BLO in those challenging scenarios, such as BLO with functional
constraints and pessimistic BLO. In this work, by reformulating BLO into
approximated single-level problems, we provide a new algorithm, named Bi-level
Value-Function-based Sequential Minimization (BVFSM), to address the above
issues. Specifically, BVFSM constructs a series of value-function-based
approximations, and thus avoids repeated calculations of recurrent gradient and
Hessian inverse required by existing approaches, time-consuming especially for
high-dimensional tasks. We also extend BVFSM to address BLO with additional
functional constraints. More importantly, BVFSM can be used for the challenging
pessimistic BLO, which has never been properly solved before. In theory, we
prove the asymptotic convergence of BVFSM on these types of BLO, in which the
restrictive lower-level convexity assumption is discarded. To our best
knowledge, this is the first gradient-based algorithm that can solve different
kinds of BLO (e.g., optimistic, pessimistic, and with constraints) with solid
convergence guarantees. Extensive experiments verify the theoretical
investigations and demonstrate our superiority on various real-world
applications.
Related papers
- A Primal-Dual-Assisted Penalty Approach to Bilevel Optimization with Coupled Constraints [66.61399765513383]
We develop a BLOCC algorithm to tackle BiLevel Optimization problems with Coupled Constraints.
We demonstrate its effectiveness on two well-known real-world applications.
arXiv Detail & Related papers (2024-06-14T15:59:36Z) - Moreau Envelope for Nonconvex Bi-Level Optimization: A Single-loop and Hessian-free Solution Strategy [45.982542530484274]
Large-scale non Bi-Level (BLO) problems are increasingly applied in machine learning.
These challenges involve ensuring computational efficiency and providing theoretical guarantees.
arXiv Detail & Related papers (2024-05-16T09:33:28Z) - Constrained Bi-Level Optimization: Proximal Lagrangian Value function
Approach and Hessian-free Algorithm [8.479947546216131]
We develop a Hessian-free gradient-based algorithm-termed proximal Lagrangian Value function-based Hessian-free Bi-level Algorithm (LV-HBA)
LV-HBA is especially well-suited for machine learning applications.
arXiv Detail & Related papers (2024-01-29T13:50:56Z) - An Introduction to Bi-level Optimization: Foundations and Applications
in Signal Processing and Machine Learning [46.02026158913706]
Bi-level optimization (BLO) has taken center stage in some exciting developments in the area of signal processing (SP) and machine learning (ML)
BLO is a classical optimization problem that involves two levels of hierarchy (i.e., upper and lower levels)
Prominent applications of BLO range from resource allocation for wireless systems to adversarial machine learning.
arXiv Detail & Related papers (2023-08-01T18:59:07Z) - Averaged Method of Multipliers for Bi-Level Optimization without
Lower-Level Strong Convexity [43.092883358935545]
We propose a single loop Bi-level Averaged Method of Multipliers (sl-BAMM) for Bi-Level Optimization (BLO)
We provide non-asymptotic convergence analysis of sl-BAMM towards KKT stationary points, and the comparative advantage of our analysis lies in the absence of strong gradient boundedness assumption.
arXiv Detail & Related papers (2023-02-07T11:29:05Z) - Enhanced Bilevel Optimization via Bregman Distance [104.96004056928474]
We propose a bilevel optimization method based on Bregman Bregman functions.
We also propose an accelerated version of SBiO-BreD method (ASBiO-BreD) by using the variance-reduced technique.
arXiv Detail & Related papers (2021-07-26T16:18:43Z) - A Generic Descent Aggregation Framework for Gradient-based Bi-level
Optimization [41.894281911990554]
We develop a novel Bi-level Descent Aggregation (BDA) framework for bi-level learning tasks.
BDA aggregates hierarchical objectives of both upper level and lower level.
We propose a new proof recipe to improve the convergence results of conventional gradient-based bi-level methods.
arXiv Detail & Related papers (2021-02-16T06:58:12Z) - A Generic First-Order Algorithmic Framework for Bi-Level Programming
Beyond Lower-Level Singleton [49.23948907229656]
Bi-level Descent Aggregation is a flexible and modularized algorithmic framework for generic bi-level optimization.
We derive a new methodology to prove the convergence of BDA without the LLS condition.
Our investigations also demonstrate that BDA is indeed compatible to a verify of particular first-order computation modules.
arXiv Detail & Related papers (2020-06-07T05:18:50Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.