Decentralized Local Stochastic Extra-Gradient for Variational
Inequalities
- URL: http://arxiv.org/abs/2106.08315v3
- Date: Sun, 2 Apr 2023 10:30:33 GMT
- Title: Decentralized Local Stochastic Extra-Gradient for Variational
Inequalities
- Authors: Aleksandr Beznosikov, Pavel Dvurechensky, Anastasia Koloskova,
Valentin Samokhin, Sebastian U Stich, Alexander Gasnikov
- Abstract summary: We consider distributed variational inequalities (VIs) on domains with the problem data that is heterogeneous (non-IID) and distributed across many devices.
We make a very general assumption on the computational network that covers the settings of fully decentralized calculations.
We theoretically analyze its convergence rate in the strongly-monotone, monotone, and non-monotone settings.
- Score: 125.62877849447729
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We consider distributed stochastic variational inequalities (VIs) on
unbounded domains with the problem data that is heterogeneous (non-IID) and
distributed across many devices. We make a very general assumption on the
computational network that, in particular, covers the settings of fully
decentralized calculations with time-varying networks and centralized
topologies commonly used in Federated Learning. Moreover, multiple local
updates on the workers can be made for reducing the communication frequency
between the workers. We extend the stochastic extragradient method to this very
general setting and theoretically analyze its convergence rate in the
strongly-monotone, monotone, and non-monotone (when a Minty solution exists)
settings. The provided rates explicitly exhibit the dependence on network
characteristics (e.g., mixing time), iteration counter, data heterogeneity,
variance, number of devices, and other standard parameters. As a special case,
our method and analysis apply to distributed stochastic saddle-point problems
(SPP), e.g., to the training of Deep Generative Adversarial Networks (GANs) for
which decentralized training has been reported to be extremely challenging. In
experiments for the decentralized training of GANs we demonstrate the
effectiveness of our proposed approach.
Related papers
- Collaborative Heterogeneous Causal Inference Beyond Meta-analysis [68.4474531911361]
We propose a collaborative inverse propensity score estimator for causal inference with heterogeneous data.
Our method shows significant improvements over the methods based on meta-analysis when heterogeneity increases.
arXiv Detail & Related papers (2024-04-24T09:04:36Z) - Averaging Rate Scheduler for Decentralized Learning on Heterogeneous
Data [7.573297026523597]
State-of-the-art decentralized learning algorithms typically require the data distribution to be Independent and Identically Distributed (IID)
We propose averaging rate scheduling as a simple yet effective way to reduce the impact of heterogeneity in decentralized learning.
arXiv Detail & Related papers (2024-03-05T19:47:51Z) - FedILC: Weighted Geometric Mean and Invariant Gradient Covariance for
Federated Learning on Non-IID Data [69.0785021613868]
Federated learning is a distributed machine learning approach which enables a shared server model to learn by aggregating the locally-computed parameter updates with the training data from spatially-distributed client silos.
We propose the Federated Invariant Learning Consistency (FedILC) approach, which leverages the gradient covariance and the geometric mean of Hessians to capture both inter-silo and intra-silo consistencies.
This is relevant to various fields such as medical healthcare, computer vision, and the Internet of Things (IoT)
arXiv Detail & Related papers (2022-05-19T03:32:03Z) - Optimal Algorithms for Decentralized Stochastic Variational Inequalities [113.43047601775453]
This work concentrates on the decentralized setting, which is increasingly important but not well understood.
We present lower bounds for both communication and local iterations and construct optimal algorithms that match these lower bounds.
Our algorithms are the best among the available not only in the decentralized case, but also in the deterministic and non-distributed literature.
arXiv Detail & Related papers (2022-02-06T13:14:02Z) - Finite-Time Convergence Rates of Decentralized Stochastic Approximation
with Applications in Multi-Agent and Multi-Task Learning [16.09467599829253]
We study a data-driven approach for finding the root of an operator under noisy measurements.
A network of agents, each with its own operator and data observations, cooperatively find the fixed point of the aggregate operator over a decentralized communication graph.
Our main contribution is to provide a finite-time analysis of this decentralized approximation method when the data observed at each agent are sampled from a Markov process.
arXiv Detail & Related papers (2020-10-28T17:01:54Z) - Generalization Error for Linear Regression under Distributed Learning [0.0]
We consider the setting where the unknowns are distributed over a network of nodes.
We present an analytical characterization of the dependence of the generalization error on the partitioning of the unknowns over nodes.
arXiv Detail & Related papers (2020-04-30T08:49:46Z) - A Unified Theory of Decentralized SGD with Changing Topology and Local
Updates [70.9701218475002]
We introduce a unified convergence analysis of decentralized communication methods.
We derive universal convergence rates for several applications.
Our proofs rely on weak assumptions.
arXiv Detail & Related papers (2020-03-23T17:49:15Z) - When Relation Networks meet GANs: Relation GANs with Triplet Loss [110.7572918636599]
Training stability is still a lingering concern of generative adversarial networks (GANs)
In this paper, we explore a relation network architecture for the discriminator and design a triplet loss which performs better generalization and stability.
Experiments on benchmark datasets show that the proposed relation discriminator and new loss can provide significant improvement on variable vision tasks.
arXiv Detail & Related papers (2020-02-24T11:35:28Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.