Invertible Monotone Operators for Normalizing Flows
- URL: http://arxiv.org/abs/2210.08176v1
- Date: Sat, 15 Oct 2022 03:40:46 GMT
- Title: Invertible Monotone Operators for Normalizing Flows
- Authors: Byeongkeun Ahn, Chiyoon Kim, Youngjoon Hong, Hyunwoo J. Kim
- Abstract summary: Normalizing flows model probability by learning invertible transformations that transfer a simple distribution into complex distributions.
We propose the monotone formulation to overcome the issue of the Lipschitz constants using monotone operators.
The resulting model, Monotone Flows, exhibits an excellent performance on multiple density estimation benchmarks.
- Score: 7.971699294672282
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Normalizing flows model probability distributions by learning invertible
transformations that transfer a simple distribution into complex distributions.
Since the architecture of ResNet-based normalizing flows is more flexible than
that of coupling-based models, ResNet-based normalizing flows have been widely
studied in recent years. Despite their architectural flexibility, it is
well-known that the current ResNet-based models suffer from constrained
Lipschitz constants. In this paper, we propose the monotone formulation to
overcome the issue of the Lipschitz constants using monotone operators and
provide an in-depth theoretical analysis. Furthermore, we construct an
activation function called Concatenated Pila (CPila) to improve gradient flow.
The resulting model, Monotone Flows, exhibits an excellent performance on
multiple density estimation benchmarks (MNIST, CIFAR-10, ImageNet32,
ImageNet64). Code is available at https://github.com/mlvlab/MonotoneFlows.
Related papers
- Free-form Flows: Make Any Architecture a Normalizing Flow [8.163244519983298]
We develop a training procedure that uses an efficient estimator for the gradient of the change of variables formula.
This enables any dimension-preserving neural network to serve as a generative model through maximum likelihood training.
We achieve excellent results in molecule generation benchmarks utilizing $E(n)$-equivariant networks.
arXiv Detail & Related papers (2023-10-25T13:23:08Z) - Locality-constrained autoregressive cum conditional normalizing flow for
lattice field theory simulations [0.0]
Local action integral leads to simplifications to the input domain of conditional normalizing flows.
We find that the autocorrelation times of l-ACNF models outperform an equivalent normalizing flow model on the full lattice by orders of magnitude.
arXiv Detail & Related papers (2023-04-04T13:55:51Z) - Lipschitz Continuity Retained Binary Neural Network [52.17734681659175]
We introduce the Lipschitz continuity as the rigorous criteria to define the model robustness for BNN.
We then propose to retain the Lipschitz continuity as a regularization term to improve the model robustness.
Our experiments prove that our BNN-specific regularization method can effectively strengthen the robustness of BNN.
arXiv Detail & Related papers (2022-07-13T22:55:04Z) - Stochastic normalizing flows as non-equilibrium transformations [62.997667081978825]
We show that normalizing flows provide a route to sample lattice field theories more efficiently than conventional MonteCarlo simulations.
We lay out a strategy to optimize the efficiency of this extended class of generative models and present examples of applications.
arXiv Detail & Related papers (2022-01-21T19:00:18Z) - Robust and Provably Monotonic Networks [0.0]
We present a new method to constrain the Lipschitz constant of dense deep learning models.
We show how the algorithm was used to train a powerful, robust, and interpretable discriminator for heavy-flavor decays in the LHCb realtime data-processing system.
arXiv Detail & Related papers (2021-11-30T19:01:32Z) - Attentive Contractive Flow with Lipschitz-constrained Self-Attention [25.84621883831624]
We introduce a novel approach called Attentive Contractive Flow (ACF)
ACF utilizes a special category of flow-based generative models - contractive flows.
We demonstrate that ACF can be introduced into a variety of state of the art flow models in a plug-and-play manner.
arXiv Detail & Related papers (2021-09-24T18:02:49Z) - Stabilizing Equilibrium Models by Jacobian Regularization [151.78151873928027]
Deep equilibrium networks (DEQs) are a new class of models that eschews traditional depth in favor of finding the fixed point of a single nonlinear layer.
We propose a regularization scheme for DEQ models that explicitly regularizes the Jacobian of the fixed-point update equations to stabilize the learning of equilibrium models.
We show that this regularization adds only minimal computational cost, significantly stabilizes the fixed-point convergence in both forward and backward passes, and scales well to high-dimensional, realistic domains.
arXiv Detail & Related papers (2021-06-28T00:14:11Z) - Robust Implicit Networks via Non-Euclidean Contractions [63.91638306025768]
Implicit neural networks show improved accuracy and significant reduction in memory consumption.
They can suffer from ill-posedness and convergence instability.
This paper provides a new framework to design well-posed and robust implicit neural networks.
arXiv Detail & Related papers (2021-06-06T18:05:02Z) - Invertible DenseNets with Concatenated LipSwish [8.315801422499861]
We introduce Invertible Dense Networks (i-DenseNets) as a more efficient alternative to Residual Flows.
We show that the proposed model out-performs Residual Flows when trained as a hybrid model where the model is both a generative and a discriminative model.
arXiv Detail & Related papers (2021-02-04T15:45:33Z) - SurVAE Flows: Surjections to Bridge the Gap between VAEs and Flows [78.77808270452974]
SurVAE Flows is a modular framework for composable transformations that encompasses VAEs and normalizing flows.
We show that several recently proposed methods, including dequantization and augmented normalizing flows, can be expressed as SurVAE Flows.
arXiv Detail & Related papers (2020-07-06T13:13:22Z) - Monotone operator equilibrium networks [97.86610752856987]
We develop a new class of implicit-depth model based on the theory of monotone operators, the Monotone Operator Equilibrium Network (monDEQ)
We show the close connection between finding the equilibrium point of an implicit network and solving a form of monotone operator splitting problem.
We then develop a parameterization of the network which ensures that all operators remain monotone, which guarantees the existence of a unique equilibrium point.
arXiv Detail & Related papers (2020-06-15T17:57:31Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.