Adaptive Multi-stage Density Ratio Estimation for Learning Latent Space
Energy-based Model
- URL: http://arxiv.org/abs/2209.08739v1
- Date: Mon, 19 Sep 2022 03:20:15 GMT
- Title: Adaptive Multi-stage Density Ratio Estimation for Learning Latent Space
Energy-based Model
- Authors: Zhisheng Xiao, Tian Han
- Abstract summary: This paper studies the fundamental problem of learning energy-based model (EBM) in the latent space of the generator model.
We propose to use noise contrastive estimation (NCE) to discriminatively learn the EBM through density ratio estimation.
Our experiments demonstrate strong performances in image generation and reconstruction as well as anomaly detection.
- Score: 11.401637217963511
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper studies the fundamental problem of learning energy-based model
(EBM) in the latent space of the generator model. Learning such prior model
typically requires running costly Markov Chain Monte Carlo (MCMC). Instead, we
propose to use noise contrastive estimation (NCE) to discriminatively learn the
EBM through density ratio estimation between the latent prior density and
latent posterior density. However, the NCE typically fails to accurately
estimate such density ratio given large gap between two densities. To
effectively tackle this issue and learn more expressive prior models, we
develop the adaptive multi-stage density ratio estimation which breaks the
estimation into multiple stages and learn different stages of density ratio
sequentially and adaptively. The latent prior model can be gradually learned
using ratio estimated in previous stage so that the final latent space EBM
prior can be naturally formed by product of ratios in different stages. The
proposed method enables informative and much sharper prior than existing
baselines, and can be trained efficiently. Our experiments demonstrate strong
performances in image generation and reconstruction as well as anomaly
detection.
Related papers
- Overcoming Saturation in Density Ratio Estimation by Iterated Regularization [11.244546184962996]
We show that a class of kernel methods for density ratio estimation suffers from error saturation.
We introduce iterated regularization in density ratio estimation to achieve fast error rates.
arXiv Detail & Related papers (2024-02-21T16:02:14Z) - Learning Energy-Based Prior Model with Diffusion-Amortized MCMC [89.95629196907082]
Common practice of learning latent space EBMs with non-convergent short-run MCMC for prior and posterior sampling is hindering the model from further progress.
We introduce a simple but effective diffusion-based amortization method for long-run MCMC sampling and develop a novel learning algorithm for the latent space EBM based on it.
arXiv Detail & Related papers (2023-10-05T00:23:34Z) - Learning Energy-Based Models by Cooperative Diffusion Recovery Likelihood [64.95663299945171]
Training energy-based models (EBMs) on high-dimensional data can be both challenging and time-consuming.
There exists a noticeable gap in sample quality between EBMs and other generative frameworks like GANs and diffusion models.
We propose cooperative diffusion recovery likelihood (CDRL), an effective approach to tractably learn and sample from a series of EBMs.
arXiv Detail & Related papers (2023-09-10T22:05:24Z) - Balanced Training of Energy-Based Models with Adaptive Flow Sampling [13.951904929884618]
Energy-based models (EBMs) are versatile density estimation models that directly parameterize an unnormalized log density.
We propose a new maximum likelihood training algorithm for EBMs that uses a different type of generative model, normalizing flows (NF)
Our method fits an NF to an EBM during training so that an NF-assisted sampling scheme provides an accurate gradient for the EBMs at all times.
arXiv Detail & Related papers (2023-06-01T13:58:06Z) - How Much is Enough? A Study on Diffusion Times in Score-based Generative
Models [76.76860707897413]
Current best practice advocates for a large T to ensure that the forward dynamics brings the diffusion sufficiently close to a known and simple noise distribution.
We show how an auxiliary model can be used to bridge the gap between the ideal and the simulated forward dynamics, followed by a standard reverse diffusion process.
arXiv Detail & Related papers (2022-06-10T15:09:46Z) - Out-of-Distribution Detection with Class Ratio Estimation [4.930817402876787]
Density-based Out-of-distribution (OOD) detection has recently been shown unreliable for the task of detecting OOD images.
We propose to unify density ratio based methods under a novel framework that builds energy-based models and employs differing base distributions.
arXiv Detail & Related papers (2022-06-08T15:20:49Z) - Posterior Coreset Construction with Kernelized Stein Discrepancy for
Model-Based Reinforcement Learning [78.30395044401321]
We develop a novel model-based approach to reinforcement learning (MBRL)
It relaxes the assumptions on the target transition model to belong to a generic family of mixture models.
It can achieve up-to 50 percent reduction in wall clock time in some continuous control environments.
arXiv Detail & Related papers (2022-06-02T17:27:49Z) - An Energy-Based Prior for Generative Saliency [62.79775297611203]
We propose a novel generative saliency prediction framework that adopts an informative energy-based model as a prior distribution.
With the generative saliency model, we can obtain a pixel-wise uncertainty map from an image, indicating model confidence in the saliency prediction.
Experimental results show that our generative saliency model with an energy-based prior can achieve not only accurate saliency predictions but also reliable uncertainty maps consistent with human perception.
arXiv Detail & Related papers (2022-04-19T10:51:00Z) - Imitation with Neural Density Models [98.34503611309256]
We propose a new framework for Imitation Learning (IL) via density estimation of the expert's occupancy measure followed by Imitation Occupancy Entropy Reinforcement Learning (RL) using the density as a reward.
Our approach maximizes a non-adversarial model-free RL objective that provably lower bounds reverse Kullback-Leibler divergence between occupancy measures of the expert and imitator.
arXiv Detail & Related papers (2020-10-19T19:38:36Z) - Telescoping Density-Ratio Estimation [21.514983459970903]
We introduce a new framework, telescoping density-ratio estimation (TRE)
TRE enables the estimation of ratios between highly dissimilar densities in high-dimensional spaces.
Our experiments demonstrate that TRE can yield substantial improvements over existing single-ratio methods.
arXiv Detail & Related papers (2020-06-22T12:55:06Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.