Learning Latent Space Energy-Based Prior Model
- URL: http://arxiv.org/abs/2006.08205v2
- Date: Thu, 29 Oct 2020 12:35:01 GMT
- Title: Learning Latent Space Energy-Based Prior Model
- Authors: Bo Pang, Tian Han, Erik Nijkamp, Song-Chun Zhu, Ying Nian Wu
- Abstract summary: We learn energy-based model (EBM) in the latent space of a generator model.
We show that the learned model exhibits strong performances in terms of image and text generation and anomaly detection.
- Score: 118.86447805707094
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We propose to learn energy-based model (EBM) in the latent space of a
generator model, so that the EBM serves as a prior model that stands on the
top-down network of the generator model. Both the latent space EBM and the
top-down network can be learned jointly by maximum likelihood, which involves
short-run MCMC sampling from both the prior and posterior distributions of the
latent vector. Due to the low dimensionality of the latent space and the
expressiveness of the top-down network, a simple EBM in latent space can
capture regularities in the data effectively, and MCMC sampling in latent space
is efficient and mixes well. We show that the learned model exhibits strong
performances in terms of image and text generation and anomaly detection. The
one-page code can be found in supplementary materials.
Related papers
- Learning Latent Space Hierarchical EBM Diffusion Models [4.4996462447311725]
We study the learning problem of the energy-based prior model and the multi-layer generator model.
Recent works have explored learning the energy-based (EBM) prior model as a second-stage, complementary model to bridge the gap.
We propose to leverage the diffusion probabilistic scheme to mitigate the burden of EBM sampling and thus facilitate EBM learning.
arXiv Detail & Related papers (2024-05-22T18:34:25Z) - Learning Energy-based Model via Dual-MCMC Teaching [5.31573596283377]
Learning the energy-based model (EBM) can be achieved using the maximum likelihood estimation (MLE)
This paper studies the fundamental learning problem of the energy-based model (EBM)
arXiv Detail & Related papers (2023-12-05T03:39:54Z) - Learning Energy-Based Prior Model with Diffusion-Amortized MCMC [89.95629196907082]
Common practice of learning latent space EBMs with non-convergent short-run MCMC for prior and posterior sampling is hindering the model from further progress.
We introduce a simple but effective diffusion-based amortization method for long-run MCMC sampling and develop a novel learning algorithm for the latent space EBM based on it.
arXiv Detail & Related papers (2023-10-05T00:23:34Z) - Learning Energy-Based Models by Cooperative Diffusion Recovery Likelihood [64.95663299945171]
Training energy-based models (EBMs) on high-dimensional data can be both challenging and time-consuming.
There exists a noticeable gap in sample quality between EBMs and other generative frameworks like GANs and diffusion models.
We propose cooperative diffusion recovery likelihood (CDRL), an effective approach to tractably learn and sample from a series of EBMs.
arXiv Detail & Related papers (2023-09-10T22:05:24Z) - Latent Diffusion Energy-Based Model for Interpretable Text Modeling [104.85356157724372]
We introduce a novel symbiosis between the diffusion models and latent space EBMs in a variational learning framework.
We develop a geometric clustering-based regularization jointly with the information bottleneck to further improve the quality of the learned latent space.
arXiv Detail & Related papers (2022-06-13T03:41:31Z) - No MCMC for me: Amortized sampling for fast and stable training of
energy-based models [62.1234885852552]
Energy-Based Models (EBMs) present a flexible and appealing way to represent uncertainty.
We present a simple method for training EBMs at scale using an entropy-regularized generator to amortize the MCMC sampling.
Next, we apply our estimator to the recently proposed Joint Energy Model (JEM), where we match the original performance with faster and stable training.
arXiv Detail & Related papers (2020-10-08T19:17:20Z) - MCMC Should Mix: Learning Energy-Based Model with Neural Transport
Latent Space MCMC [110.02001052791353]
Learning energy-based model (EBM) requires MCMC sampling of the learned model as an inner loop of the learning algorithm.
We show that the model has a particularly simple form in the space of the latent variables of the backbone model.
arXiv Detail & Related papers (2020-06-12T01:25:51Z) - Generalized Energy Based Models [35.49065282173972]
We introduce the Generalized Energy Based Model (GEBM) for generative modelling.
Both the energy function and base jointly constitute the final model, unlike GANs, which retain only the base distribution.
We show that both training stages are well-defined.
arXiv Detail & Related papers (2020-03-10T23:22:09Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.