Convergence of GANs Training: A Game and Stochastic Control Methodology
- URL: http://arxiv.org/abs/2112.00222v1
- Date: Wed, 1 Dec 2021 01:52:23 GMT
- Title: Convergence of GANs Training: A Game and Stochastic Control Methodology
- Authors: Othmane Mounjid, Xin Guo
- Abstract summary: Training of generative adversarial networks (GANs) is known for its difficulty to converge.
This paper first confirms the lack of convexity in GANs objective functions, hence the well-posedness problem of GANs models.
In particular, it presents an optimal solution for adaptive learning rate which depends on the convexity of the objective function.
- Score: 4.933916728941277
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Training of generative adversarial networks (GANs) is known for its
difficulty to converge. This paper first confirms analytically one of the
culprits behind this convergence issue: the lack of convexity in GANs objective
functions, hence the well-posedness problem of GANs models. Then, it proposes a
stochastic control approach for hyper-parameters tuning in GANs training. In
particular, it presents an optimal solution for adaptive learning rate which
depends on the convexity of the objective function, and builds a precise
relation between improper choices of learning rate and explosion in GANs
training. Finally, empirical studies demonstrate that training algorithms
incorporating this selection methodology outperform standard ones.
Related papers
- Diffusion Models as Network Optimizers: Explorations and Analysis [71.69869025878856]
generative diffusion models (GDMs) have emerged as a promising new approach to network optimization.
In this study, we first explore the intrinsic characteristics of generative models.
We provide a concise theoretical and intuitive demonstration of the advantages of generative models over discriminative network optimization.
arXiv Detail & Related papers (2024-11-01T09:05:47Z) - Adaptive Federated Learning Over the Air [108.62635460744109]
We propose a federated version of adaptive gradient methods, particularly AdaGrad and Adam, within the framework of over-the-air model training.
Our analysis shows that the AdaGrad-based training algorithm converges to a stationary point at the rate of $mathcalO( ln(T) / T 1 - frac1alpha ).
arXiv Detail & Related papers (2024-03-11T09:10:37Z) - Error analysis of generative adversarial network [0.0]
We study the error convergence rate of the GAN model based on a class of functions encompassing the discriminator and generator neural networks.
By employing the Talagrand inequality and Borel-Cantelli lemma, we establish a tight convergence rate for the error of GAN.
arXiv Detail & Related papers (2023-10-23T22:39:28Z) - Stochastic Unrolled Federated Learning [85.6993263983062]
We introduce UnRolled Federated learning (SURF), a method that expands algorithm unrolling to federated learning.
Our proposed method tackles two challenges of this expansion, namely the need to feed whole datasets to the unrolleds and the decentralized nature of federated learning.
arXiv Detail & Related papers (2023-05-24T17:26:22Z) - Promises and Pitfalls of the Linearized Laplace in Bayesian Optimization [73.80101701431103]
The linearized-Laplace approximation (LLA) has been shown to be effective and efficient in constructing Bayesian neural networks.
We study the usefulness of the LLA in Bayesian optimization and highlight its strong performance and flexibility.
arXiv Detail & Related papers (2023-04-17T14:23:43Z) - Learning MDPs from Features: Predict-Then-Optimize for Sequential
Decision Problems by Reinforcement Learning [52.74071439183113]
We study the predict-then-optimize framework in the context of sequential decision problems (formulated as MDPs) solved via reinforcement learning.
Two significant computational challenges arise in applying decision-focused learning to MDPs.
arXiv Detail & Related papers (2021-06-06T23:53:31Z) - Decentralized Statistical Inference with Unrolled Graph Neural Networks [26.025935320024665]
We propose a learning-based framework, which unrolls decentralized optimization algorithms into graph neural networks (GNNs)
By minimizing the recovery error via end-to-end training, this learning-based framework resolves the model mismatch issue.
Our convergence analysis reveals that the learned model parameters may accelerate the convergence and reduce the recovery error to a large extent.
arXiv Detail & Related papers (2021-04-04T07:52:34Z) - Training Generative Adversarial Networks by Solving Ordinary
Differential Equations [54.23691425062034]
We study the continuous-time dynamics induced by GAN training.
From this perspective, we hypothesise that instabilities in training GANs arise from the integration error.
We experimentally verify that well-known ODE solvers (such as Runge-Kutta) can stabilise training.
arXiv Detail & Related papers (2020-10-28T15:23:49Z) - The Strength of Nesterov's Extrapolation in the Individual Convergence
of Nonsmooth Optimization [0.0]
We prove that Nesterov's extrapolation has the strength to make the individual convergence of gradient descent methods optimal for nonsmooth problems.
We give an extension of the derived algorithms to solve regularized learning tasks with nonsmooth losses in settings.
Our method is applicable as an efficient tool for solving large-scale $l$1-regularized hinge-loss learning problems.
arXiv Detail & Related papers (2020-06-08T03:35:41Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.