A Tale of Two Flows: Cooperative Learning of Langevin Flow and
Normalizing Flow Toward Energy-Based Model
- URL: http://arxiv.org/abs/2205.06924v1
- Date: Fri, 13 May 2022 23:12:38 GMT
- Title: A Tale of Two Flows: Cooperative Learning of Langevin Flow and
Normalizing Flow Toward Energy-Based Model
- Authors: Jianwen Xie, Yaxuan Zhu, Jun Li, Ping Li
- Abstract summary: We study the cooperative learning of two generative flow models, in which the two models are iteratively updated based on jointly synthesized examples.
We show that the trained CoopFlow is capable of realistic images, reconstructing images, and interpolating between images.
- Score: 43.53802699867521
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: This paper studies the cooperative learning of two generative flow models, in
which the two models are iteratively updated based on the jointly synthesized
examples. The first flow model is a normalizing flow that transforms an initial
simple density to a target density by applying a sequence of invertible
transformations. The second flow model is a Langevin flow that runs finite
steps of gradient-based MCMC toward an energy-based model. We start from
proposing a generative framework that trains an energy-based model with a
normalizing flow as an amortized sampler to initialize the MCMC chains of the
energy-based model. In each learning iteration, we generate synthesized
examples by using a normalizing flow initialization followed by a short-run
Langevin flow revision toward the current energy-based model. Then we treat the
synthesized examples as fair samples from the energy-based model and update the
model parameters with the maximum likelihood learning gradient, while the
normalizing flow directly learns from the synthesized examples by maximizing
the tractable likelihood. Under the short-run non-mixing MCMC scenario, the
estimation of the energy-based model is shown to follow the perturbation of
maximum likelihood, and the short-run Langevin flow and the normalizing flow
form a two-flow generator that we call CoopFlow. We provide an understating of
the CoopFlow algorithm by information geometry and show that it is a valid
generator as it converges to a moment matching estimator. We demonstrate that
the trained CoopFlow is capable of synthesizing realistic images,
reconstructing images, and interpolating between images.
Related papers
- Latent Space Energy-based Neural ODEs [73.01344439786524]
This paper introduces a novel family of deep dynamical models designed to represent continuous-time sequence data.
We train the model using maximum likelihood estimation with Markov chain Monte Carlo.
Experiments on oscillating systems, videos and real-world state sequences (MuJoCo) illustrate that ODEs with the learnable energy-based prior outperform existing counterparts.
arXiv Detail & Related papers (2024-09-05T18:14:22Z) - Variational Potential Flow: A Novel Probabilistic Framework for Energy-Based Generative Modelling [10.926841288976684]
We present a novel energy-based generative framework, Variational Potential Flow (VAPO)
VAPO aims to learn a potential energy function whose gradient (flow) guides the prior samples, so that their density evolution closely follows an approximate data likelihood homotopy.
Images can be generated after training the potential energy, by initializing the samples from Gaussian prior and solving the ODE governing the potential flow on a fixed time interval.
arXiv Detail & Related papers (2024-07-21T18:08:12Z) - Convolutional L2LFlows: Generating Accurate Showers in Highly Granular Calorimeters Using Convolutional Normalizing Flows [0.0]
We extend L2LFlows to simulate showers with a 9-times larger profile in the lateral direction.
We introduce convolutional layers and U-Net-type connections, move from masked autoregressive flows to coupling layers.
arXiv Detail & Related papers (2024-05-30T18:25:19Z) - Guided Flows for Generative Modeling and Decision Making [55.42634941614435]
We show that Guided Flows significantly improves the sample quality in conditional image generation and zero-shot text synthesis-to-speech.
Notably, we are first to apply flow models for plan generation in the offline reinforcement learning setting ax speedup in compared to diffusion models.
arXiv Detail & Related papers (2023-11-22T15:07:59Z) - A Tale of Two Latent Flows: Learning Latent Space Normalizing Flow with
Short-run Langevin Flow for Approximate Inference [44.97938062814525]
We study a normalizing flow in the latent space of a top-down generator model, in which the normalizing flow model plays the role of the informative prior model of the generator.
We propose to jointly learn the latent space normalizing flow prior model and the top-down generator model by a Markov chain Monte Carlo (MCMC)-based maximum likelihood algorithm.
arXiv Detail & Related papers (2023-01-23T07:15:43Z) - Generative Flow Networks for Discrete Probabilistic Modeling [118.81967600750428]
We present energy-based generative flow networks (EB-GFN)
EB-GFN is a novel probabilistic modeling algorithm for high-dimensional discrete data.
We show how GFlowNets can approximately perform large-block Gibbs sampling to mix between modes.
arXiv Detail & Related papers (2022-02-03T01:27:11Z) - CaloFlow: Fast and Accurate Generation of Calorimeter Showers with
Normalizing Flows [0.0]
We introduce CaloFlow, a fast detector simulation framework based on normalizing flows.
For the first time, we demonstrate that normalizing flows can reproduce many-channel calorimeter showers with extremely high fidelity.
arXiv Detail & Related papers (2021-06-09T18:00:02Z) - Closing the Dequantization Gap: PixelCNN as a Single-Layer Flow [16.41460104376002]
We introduce subset flows, a class of flows that can transform finite volumes and allow exact computation of likelihoods for discrete data.
We identify ordinal discrete autoregressive models, including WaveNets, PixelCNNs and Transformers, as single-layer flows.
We demonstrate state-of-the-art results on CIFAR-10 for flow models trained with dequantization.
arXiv Detail & Related papers (2020-02-06T22:58:51Z) - A Near-Optimal Gradient Flow for Learning Neural Energy-Based Models [93.24030378630175]
We propose a novel numerical scheme to optimize the gradient flows for learning energy-based models (EBMs)
We derive a second-order Wasserstein gradient flow of the global relative entropy from Fokker-Planck equation.
Compared with existing schemes, Wasserstein gradient flow is a smoother and near-optimal numerical scheme to approximate real data densities.
arXiv Detail & Related papers (2019-10-31T02:26:20Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.