Optimal Algorithms for Convex Nested Stochastic Composite Optimization
- URL: http://arxiv.org/abs/2011.10076v5
- Date: Tue, 21 Jun 2022 15:53:19 GMT
- Title: Optimal Algorithms for Convex Nested Stochastic Composite Optimization
- Authors: Zhe Zhang, Guanghui Lan
- Abstract summary: convex nested composite optimization (NSCO) has received considerable attention for its applications in reinforcement learning and risk-averse optimization.
The current NSCO algorithms have worse oracle complexities, by orders of magnitude, than those for simpler composite optimization problems without the nested structure.
We develop order-optimal algorithms for the convex NSCO problem constructed from an arbitrary composition of smooth, structured non-smooth and general non-smooth layer functions.
- Score: 13.200502573462712
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Recently, convex nested stochastic composite optimization (NSCO) has received
considerable attention for its applications in reinforcement learning and
risk-averse optimization. The current NSCO algorithms have worse stochastic
oracle complexities, by orders of magnitude, than those for simpler stochastic
composite optimization problems (e.g., sum of smooth and nonsmooth functions)
without the nested structure. Moreover, they require all outer-layer functions
to be smooth, which is not satisfied by some important applications. These
discrepancies prompt us to ask: ``does the nested composition make stochastic
optimization more difficult in terms of the order of oracle complexity?" In
this paper, we answer the question by developing order-optimal algorithms for
the convex NSCO problem constructed from an arbitrary composition of smooth,
structured non-smooth and general non-smooth layer functions. When all
outer-layer functions are smooth, we propose a stochastic sequential dual (SSD)
method to achieve an oracle complexity of $\mathcal{O}(1/\epsilon^2)$
($\mathcal{O}(1/\epsilon)$) when the problem is non-strongly (strongly) convex.
When there exists some structured non-smooth or general non-smooth outer-layer
function, we propose a nonsmooth stochastic sequential dual (nSSD) method to
achieve an oracle complexity of $\mathcal{O}(1/\epsilon^2)$. We provide a lower
complexity bound to show the latter $\mathcal{O}(1/\epsilon^2)$ complexity to
be unimprovable even under a strongly convex setting. All these complexity
results seem to be new in the literature and they indicate that the convex NSCO
problem has the same order of oracle complexity as those without the nested
composition in all but the strongly convex and outer-non-smooth problem.
Related papers
- Stochastic Inexact Augmented Lagrangian Method for Nonconvex Expectation
Constrained Optimization [88.0031283949404]
Many real-world problems have complicated non functional constraints and use a large number of data points.
Our proposed method outperforms an existing method with the previously best-known result.
arXiv Detail & Related papers (2022-12-19T14:48:54Z) - Optimal Algorithms for Stochastic Complementary Composite Minimization [55.26935605535377]
Inspired by regularization techniques in statistics and machine learning, we study complementary composite minimization.
We provide novel excess risk bounds, both in expectation and with high probability.
Our algorithms are nearly optimal, which we prove via novel lower complexity bounds for this class of problems.
arXiv Detail & Related papers (2022-11-03T12:40:24Z) - Optimal Algorithms for Stochastic Multi-Level Compositional Optimization [46.77664277596764]
We solve the problem of multi-level compositional optimization, where the objective function is a limitation of multiple but possibly non-optimal functions.
To make use of the Adaptive Multi-level Variance Reduction method (SMVR), we also achieves the same optimal complexities but converges faster in practice.
arXiv Detail & Related papers (2022-02-15T16:02:32Z) - Randomized Stochastic Variance-Reduced Methods for Stochastic Bilevel
Optimization [62.87181271021217]
We consider non-SBO problems that have many applications in machine learning.
This paper proposes fast randomized algorithms for non-SBO problems.
arXiv Detail & Related papers (2021-05-05T18:28:42Z) - Oracle Complexity in Nonsmooth Nonconvex Optimization [49.088972349825085]
It is well-known that given a smooth, bounded-from-below $$stationary points, Oracle-based methods can find smooth approximation of smoothness.
In this paper, we prove an inherent trade-off between optimization and smoothing dimension.
arXiv Detail & Related papers (2021-04-14T10:42:45Z) - Distributed Stochastic Consensus Optimization with Momentum for
Nonconvex Nonsmooth Problems [45.88640334610308]
This paper presents a new distributed optimization algorithm for non-smooth problems.
We show that the proposed algorithm can achieve an overcal communication.
Experiments are presented to illustrate the effectiveness of the proposed algorithms.
arXiv Detail & Related papers (2020-11-10T13:12:21Z) - Stochastic Recursive Variance Reduction for Efficient Smooth Non-Convex
Compositional Optimization [20.410012564838933]
compositional optimization arises in many important machine learning tasks such as value function evaluation in reinforcement learning and portfolio management.
In this paper, we investigate the general compositional optimization in the general smooth non-cursive setting.
arXiv Detail & Related papers (2019-12-31T18:59:13Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.