Stochastic Compositional Optimization with Compositional Constraints
- URL: http://arxiv.org/abs/2209.04086v1
- Date: Fri, 9 Sep 2022 02:06:35 GMT
- Title: Stochastic Compositional Optimization with Compositional Constraints
- Authors: Shuoguang Yang, Zhe Zhang, Ethan X. Fang
- Abstract summary: We study a novel model that incorporates single-level expected value and two-level compositional constraints into the current SCO framework.
Our model can be applied widely to data-driven optimization and risk management, including risk-averse optimization and high-moment portfolio selection.
We propose a class of primal-dual algorithms that generates sequences converging to the optimal solution at the rate of $cO(frac1sqrtN)$under both single-level expected value and two-level compositional constraints.
- Score: 8.535822787583486
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Stochastic compositional optimization (SCO) has attracted considerable
attention because of its broad applicability to important real-world problems.
However, existing works on SCO assume that the projection within a solution
update is simple, which fails to hold for problem instances where the
constraints are in the form of expectations, such as empirical conditional
value-at-risk constraints. We study a novel model that incorporates
single-level expected value and two-level compositional constraints into the
current SCO framework. Our model can be applied widely to data-driven
optimization and risk management, including risk-averse optimization and
high-moment portfolio selection, and can handle multiple constraints. We
further propose a class of primal-dual algorithms that generates sequences
converging to the optimal solution at the rate of
$\cO(\frac{1}{\sqrt{N}})$under both single-level expected value and two-level
compositional constraints, where $N$ is the iteration counter, establishing the
benchmarks in expected value constrained SCO.
Related papers
- A single-loop SPIDER-type stochastic subgradient method for expectation-constrained nonconvex nonsmooth optimization [17.25924791071807]
We present a novel type of subgradient algorithm for complex constraints.
We show that our method is significantly faster than two-of-the-art algorithms.
arXiv Detail & Related papers (2025-01-31T15:18:52Z) - Optimal Rates for Robust Stochastic Convex Optimization [12.620782629498812]
We develop novel algorithms that achieve minimax-optimal excess risk under the $epsilon$-contamination model.
Our algorithms do not require restrictive assumptions, and can handle nonsmooth but Lipschitz population loss functions.
arXiv Detail & Related papers (2024-12-15T00:52:08Z) - Double Duality: Variational Primal-Dual Policy Optimization for
Constrained Reinforcement Learning [132.7040981721302]
We study the Constrained Convex Decision Process (MDP), where the goal is to minimize a convex functional of the visitation measure.
Design algorithms for a constrained convex MDP faces several challenges, including handling the large state space.
arXiv Detail & Related papers (2024-02-16T16:35:18Z) - Algorithm for Constrained Markov Decision Process with Linear
Convergence [55.41644538483948]
An agent aims to maximize the expected accumulated discounted reward subject to multiple constraints on its costs.
A new dual approach is proposed with the integration of two ingredients: entropy regularized policy and Vaidya's dual.
The proposed approach is shown to converge (with linear rate) to the global optimum.
arXiv Detail & Related papers (2022-06-03T16:26:38Z) - Online Learning with Knapsacks: the Best of Both Worlds [54.28273783164608]
We casting online learning problems in which a decision maker wants to maximize their expected reward without violating a finite set of $m$m resource constraints.
Our framework allows the decision maker to handle its evidence flexibility and costoretic functions.
arXiv Detail & Related papers (2022-02-28T12:10:48Z) - A Globally Convergent Evolutionary Strategy for Stochastic Constrained
Optimization with Applications to Reinforcement Learning [0.6445605125467573]
Evolutionary strategies have been shown to achieve competing levels of performance for complex optimization problems in reinforcement learning.
Convergence guarantees for evolutionary strategies to optimize constrained problems are however lacking in the literature.
arXiv Detail & Related papers (2022-02-21T17:04:51Z) - Faster Algorithm and Sharper Analysis for Constrained Markov Decision
Process [56.55075925645864]
The problem of constrained decision process (CMDP) is investigated, where an agent aims to maximize the expected accumulated discounted reward subject to multiple constraints.
A new utilities-dual convex approach is proposed with novel integration of three ingredients: regularized policy, dual regularizer, and Nesterov's gradient descent dual.
This is the first demonstration that nonconcave CMDP problems can attain the lower bound of $mathcal O (1/epsilon)$ for all complexity optimization subject to convex constraints.
arXiv Detail & Related papers (2021-10-20T02:57:21Z) - Momentum Accelerates the Convergence of Stochastic AUPRC Maximization [80.8226518642952]
We study optimization of areas under precision-recall curves (AUPRC), which is widely used for imbalanced tasks.
We develop novel momentum methods with a better iteration of $O (1/epsilon4)$ for finding an $epsilon$stationary solution.
We also design a novel family of adaptive methods with the same complexity of $O (1/epsilon4)$, which enjoy faster convergence in practice.
arXiv Detail & Related papers (2021-07-02T16:21:52Z) - Modeling the Second Player in Distributionally Robust Optimization [90.25995710696425]
We argue for the use of neural generative models to characterize the worst-case distribution.
This approach poses a number of implementation and optimization challenges.
We find that the proposed approach yields models that are more robust than comparable baselines.
arXiv Detail & Related papers (2021-03-18T14:26:26Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.