GAT-GMM: Generative Adversarial Training for Gaussian Mixture Models
- URL: http://arxiv.org/abs/2006.10293v1
- Date: Thu, 18 Jun 2020 06:11:28 GMT
- Title: GAT-GMM: Generative Adversarial Training for Gaussian Mixture Models
- Authors: Farzan Farnia, William Wang, Subhro Das, Ali Jadbabaie
- Abstract summary: Generative adversarial networks (GANs) learn the distribution of observed samples through a zero-sum game.
We propose Gene Adversarial Gaussian Models (GAT-GMM), a minimax GANMs.
We show that GAT-GMM can perform as well as the expectation-maximization algorithm in learning mixtures of two Gaussians.
- Score: 29.42264360774606
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Generative adversarial networks (GANs) learn the distribution of observed
samples through a zero-sum game between two machine players, a generator and a
discriminator. While GANs achieve great success in learning the complex
distribution of image, sound, and text data, they perform suboptimally in
learning multi-modal distribution-learning benchmarks including Gaussian
mixture models (GMMs). In this paper, we propose Generative Adversarial
Training for Gaussian Mixture Models (GAT-GMM), a minimax GAN framework for
learning GMMs. Motivated by optimal transport theory, we design the zero-sum
game in GAT-GMM using a random linear generator and a softmax-based quadratic
discriminator architecture, which leads to a non-convex concave minimax
optimization problem. We show that a Gradient Descent Ascent (GDA) method
converges to an approximate stationary minimax point of the GAT-GMM
optimization problem. In the benchmark case of a mixture of two symmetric,
well-separated Gaussians, we further show this stationary point recovers the
true parameters of the underlying GMM. We numerically support our theoretical
findings by performing several experiments, which demonstrate that GAT-GMM can
perform as well as the expectation-maximization algorithm in learning mixtures
of two Gaussians.
Related papers
- Toward Global Convergence of Gradient EM for Over-Parameterized Gaussian Mixture Models [47.294535652946095]
We study the gradient Expectation-Maximization (EM) algorithm for Gaussian Mixture Models (GMM)
This is the first global convergence result for Gaussian mixtures with more than $2$ components.
arXiv Detail & Related papers (2024-06-29T16:44:29Z) - Provably Efficient Information-Directed Sampling Algorithms for Multi-Agent Reinforcement Learning [50.92957910121088]
This work designs and analyzes a novel set of algorithms for multi-agent reinforcement learning (MARL) based on the principle of information-directed sampling (IDS)
For episodic two-player zero-sum MGs, we present three sample-efficient algorithms for learning Nash equilibrium.
We extend Reg-MAIDS to multi-player general-sum MGs and prove that it can learn either the Nash equilibrium or coarse correlated equilibrium in a sample efficient manner.
arXiv Detail & Related papers (2024-04-30T06:48:56Z) - Improved DDIM Sampling with Moment Matching Gaussian Mixtures [1.450405446885067]
We propose using a Gaussian Mixture Model (GMM) as reverse transition operator ( kernel) within the Denoising Diffusion Implicit Models (DDIM) framework.
We match the first and second order central moments of the DDPM forward marginals by constraining the parameters of the GMM.
Our results suggest that using the GMM kernel leads to significant improvements in the quality of the generated samples when the number of sampling steps is small.
arXiv Detail & Related papers (2023-11-08T00:24:50Z) - An Efficient 1 Iteration Learning Algorithm for Gaussian Mixture Model
And Gaussian Mixture Embedding For Neural Network [2.261786383673667]
The new algorithm brings more robustness and simplicity than classic Expectation Maximization (EM) algorithm.
It also improves the accuracy and only take 1 iteration for learning.
arXiv Detail & Related papers (2023-08-18T10:17:59Z) - Cramer Type Distances for Learning Gaussian Mixture Models by Gradient
Descent [0.0]
As of today, few known algorithms can fit or learn Gaussian mixture models.
We propose a distance function called Sliced Cram'er 2-distance for learning general multivariate GMMs.
These features are especially useful for distributional reinforcement learning and Deep Q Networks.
arXiv Detail & Related papers (2023-07-13T13:43:02Z) - Continual Learning with Fully Probabilistic Models [70.3497683558609]
We present an approach for continual learning based on fully probabilistic (or generative) models of machine learning.
We propose a pseudo-rehearsal approach using a Gaussian Mixture Model (GMM) instance for both generator and classifier functionalities.
We show that GMR achieves state-of-the-art performance on common class-incremental learning problems at very competitive time and memory complexity.
arXiv Detail & Related papers (2021-04-19T12:26:26Z) - Image Modeling with Deep Convolutional Gaussian Mixture Models [79.0660895390689]
We present a new formulation of deep hierarchical Gaussian Mixture Models (GMMs) that is suitable for describing and generating images.
DCGMMs avoid this by a stacked architecture of multiple GMM layers, linked by convolution and pooling operations.
For generating sharp images with DCGMMs, we introduce a new gradient-based technique for sampling through non-invertible operations like convolution and pooling.
Based on the MNIST and FashionMNIST datasets, we validate the DCGMMs model by demonstrating its superiority over flat GMMs for clustering, sampling and outlier detection.
arXiv Detail & Related papers (2021-04-19T12:08:53Z) - Cauchy-Schwarz Regularized Autoencoder [68.80569889599434]
Variational autoencoders (VAE) are a powerful and widely-used class of generative models.
We introduce a new constrained objective based on the Cauchy-Schwarz divergence, which can be computed analytically for GMMs.
Our objective improves upon variational auto-encoding models in density estimation, unsupervised clustering, semi-supervised learning, and face analysis.
arXiv Detail & Related papers (2021-01-06T17:36:26Z) - Plug-And-Play Learned Gaussian-mixture Approximate Message Passing [71.74028918819046]
We propose a plug-and-play compressed sensing (CS) recovery algorithm suitable for any i.i.d. source prior.
Our algorithm builds upon Borgerding's learned AMP (LAMP), yet significantly improves it by adopting a universal denoising function within the algorithm.
Numerical evaluation shows that the L-GM-AMP algorithm achieves state-of-the-art performance without any knowledge of the source prior.
arXiv Detail & Related papers (2020-11-18T16:40:45Z) - DeepGMR: Learning Latent Gaussian Mixture Models for Registration [113.74060941036664]
Point cloud registration is a fundamental problem in 3D computer vision, graphics and robotics.
In this paper, we introduce Deep Gaussian Mixture Registration (DeepGMR), the first learning-based registration method.
Our proposed method shows favorable performance when compared with state-of-the-art geometry-based and learning-based registration methods.
arXiv Detail & Related papers (2020-08-20T17:25:16Z) - Effective Learning of a GMRF Mixture Model [8.336315962271396]
We propose restricting the GMM to a Gaussian Markov Random Field Mixture Model (GMRF-MM)
When the sparsity pattern of each matrix is known, we propose an efficient optimization method for the Maximum Likelihood Estimate (MLE) of that matrix.
We show that our "debiasing" approach outperforms GLASSO in both the single-GMRF and the GMRF-MM cases.
arXiv Detail & Related papers (2020-05-18T19:00:14Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.