An Accelerated DFO Algorithm for Finite-sum Convex Functions
- URL: http://arxiv.org/abs/2007.03311v2
- Date: Mon, 3 Aug 2020 03:55:59 GMT
- Title: An Accelerated DFO Algorithm for Finite-sum Convex Functions
- Authors: Yuwen Chen (1), Antonio Orvieto (1), Aurelien Lucchi (1) ((1) ETH
Zurich)
- Abstract summary: Derivative-free optimizationDFO) has recently gained a lot momentum in machine learning.
In this paper, we exploit the finite-sum objective functions in order to design accessible DFO algorithms.
- Score: 0.0
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Derivative-free optimization (DFO) has recently gained a lot of momentum in
machine learning, spawning interest in the community to design faster methods
for problems where gradients are not accessible. While some attention has been
given to the concept of acceleration in the DFO literature, existing stochastic
algorithms for objective functions with a finite-sum structure have not been
shown theoretically to achieve an accelerated rate of convergence. Algorithms
that use acceleration in such a setting are prone to instabilities, making it
difficult to reach convergence. In this work, we exploit the finite-sum
structure of the objective in order to design a variance-reduced DFO algorithm
that provably yields acceleration. We prove rates of convergence for both
smooth convex and strongly-convex finite-sum objective functions. Finally, we
validate our theoretical results empirically on several tasks and datasets.
Related papers
- D4FT: A Deep Learning Approach to Kohn-Sham Density Functional Theory [79.50644650795012]
We propose a deep learning approach to solve Kohn-Sham Density Functional Theory (KS-DFT)
We prove that such an approach has the same expressivity as the SCF method, yet reduces the computational complexity.
In addition, we show that our approach enables us to explore more complex neural-based wave functions.
arXiv Detail & Related papers (2023-03-01T10:38:10Z) - Accelerated First-Order Optimization under Nonlinear Constraints [73.2273449996098]
We exploit between first-order algorithms for constrained optimization and non-smooth systems to design a new class of accelerated first-order algorithms.
An important property of these algorithms is that constraints are expressed in terms of velocities instead of sparse variables.
arXiv Detail & Related papers (2023-02-01T08:50:48Z) - Generalized Gradient Flows with Provable Fixed-Time Convergence and Fast
Evasion of Non-Degenerate Saddle Points [8.452349885923507]
Gradient-based first-order convex optimization algorithms find widespread applicability in a variety of domains, including machine learning tasks.
Motivated by the recent advances in fixed-time theory of optimal time, we introduce a framework for designing accelerated optimization algorithms.
For functions that admit non-de saddle-points, we show that the time required to evade these saddle-points is uniformly bounded for all initial conditions.
arXiv Detail & Related papers (2022-12-07T16:36:23Z) - Accelerating Perturbed Stochastic Iterates in Asynchronous Lock-Free
Optimization [47.03235286622727]
We prove that our new accelerated method requires the same linear speed-up condition as the existing non-accelerated methods.
Our core algorithmic discovery is a new accelerated SVRG variant with sparse updates.
arXiv Detail & Related papers (2021-09-30T17:36:32Z) - Nesterov Accelerated ADMM for Fast Diffeomorphic Image Registration [63.15453821022452]
Recent developments in approaches based on deep learning have achieved sub-second runtimes for DiffIR.
We propose a simple iterative scheme that functionally composes intermediate non-stationary velocity fields.
We then propose a convex optimisation model that uses a regularisation term of arbitrary order to impose smoothness on these velocity fields.
arXiv Detail & Related papers (2021-09-26T19:56:45Z) - Distributed stochastic optimization with large delays [59.95552973784946]
One of the most widely used methods for solving large-scale optimization problems is distributed asynchronous gradient descent (DASGD)
We show that DASGD converges to a global optimal implementation model under same delay assumptions.
arXiv Detail & Related papers (2021-07-06T21:59:49Z) - Fast Convergence Algorithm for Analog Federated Learning [30.399830943617772]
We propose an AirComp-based FedSplit algorithm for efficient analog federated learning over wireless channels.
We prove that the proposed algorithm linearly converges to the optimal solutions under the assumption that the objective function is strongly convex and smooth.
Our algorithm is theoretically and experimentally verified to be much more robust to the ill-conditioned problems with faster convergence compared with other benchmark FL algorithms.
arXiv Detail & Related papers (2020-10-30T10:59:49Z) - Accelerated Algorithms for Convex and Non-Convex Optimization on
Manifolds [9.632674803757475]
We propose a scheme for solving convex and non- optimization problems on distance.
Our proposed algorithm adapts to the level of complexity in the objective function.
arXiv Detail & Related papers (2020-10-18T02:48:22Z) - IDEAL: Inexact DEcentralized Accelerated Augmented Lagrangian Method [64.15649345392822]
We introduce a framework for designing primal methods under the decentralized optimization setting where local functions are smooth and strongly convex.
Our approach consists of approximately solving a sequence of sub-problems induced by the accelerated augmented Lagrangian method.
When coupled with accelerated gradient descent, our framework yields a novel primal algorithm whose convergence rate is optimal and matched by recently derived lower bounds.
arXiv Detail & Related papers (2020-06-11T18:49:06Z) - Accelerating Block Coordinate Descent for Nonnegative Tensor
Factorization [22.2289974576617]
This paper is concerned with improving the empirical convergence speed of block descent algorithms for approximate nonnegative factorization (NTF)
We propose an extrapolation strategy in-between block updates and restarts (HER)
HER significantly accelerates the empirical convergence speed of most existing blockcoordinate algorithms for dense NTF, in particular for challenging computational scenarios, while requiring a negligible additional computational budget.
arXiv Detail & Related papers (2020-01-13T14:59:03Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.