FInC Flow: Fast and Invertible $k \times k$ Convolutions for Normalizing
Flows
- URL: http://arxiv.org/abs/2301.09266v1
- Date: Mon, 23 Jan 2023 04:31:03 GMT
- Title: FInC Flow: Fast and Invertible $k \times k$ Convolutions for Normalizing
Flows
- Authors: Aditya Kallappa, Sandeep Nagar, Girish Varma
- Abstract summary: Invertible convolutions have been an essential element for building expressive normalizing flow-based generative models.
We propose a $k times k$ convolutional layer and Deep Normalizing Flow architecture.
- Score: 2.156373334386171
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Invertible convolutions have been an essential element for building
expressive normalizing flow-based generative models since their introduction in
Glow. Several attempts have been made to design invertible $k \times k$
convolutions that are efficient in training and sampling passes. Though these
attempts have improved the expressivity and sampling efficiency, they severely
lagged behind Glow which used only $1 \times 1$ convolutions in terms of
sampling time. Also, many of the approaches mask a large number of parameters
of the underlying convolution, resulting in lower expressivity on a fixed
run-time budget. We propose a $k \times k$ convolutional layer and Deep
Normalizing Flow architecture which i.) has a fast parallel inversion algorithm
with running time O$(n k^2)$ ($n$ is height and width of the input image and k
is kernel size), ii.) masks the minimal amount of learnable parameters in a
layer. iii.) gives better forward pass and sampling times comparable to other
$k \times k$ convolution-based models on real-world benchmarks. We provide an
implementation of the proposed parallel algorithm for sampling using our
invertible convolutions on GPUs. Benchmarks on CIFAR-10, ImageNet, and CelebA
datasets show comparable performance to previous works regarding bits per
dimension while significantly improving the sampling time.
Related papers
- Parallel Backpropagation for Inverse of a Convolution with Application to Normalizing Flows [2.048226951354646]
Inverse of an invertible convolution is an important operation that comes up in Normalizing Flows.
We give a fast parallel backpropagation algorithm with running time $O(sqrtn)$ for a square image.
We show significantly improved sampling times with similar bits per dimension compared to previous models.
arXiv Detail & Related papers (2024-10-18T17:35:33Z) - MVSplat: Efficient 3D Gaussian Splatting from Sparse Multi-View Images [102.7646120414055]
We introduce MVSplat, an efficient model that, given sparse multi-view images as input, predicts clean feed-forward 3D Gaussians.
On the large-scale RealEstate10K and ACID benchmarks, MVSplat achieves state-of-the-art performance with the fastest feed-forward inference speed (22fps)
arXiv Detail & Related papers (2024-03-21T17:59:58Z) - FlashFFTConv: Efficient Convolutions for Long Sequences with Tensor
Cores [18.016204763652553]
Convolution models with long filters have demonstrated state-of-the-art reasoning abilities in many long-sequence tasks.
Fast Fourier Transform (FFT) allows long convolutions to run in $O(N logN)$ time in sequence length $N$ but has poor hardware utilization.
In this paper, we study how to optimize the FFT convolution.
arXiv Detail & Related papers (2023-11-10T07:33:35Z) - Multi-block-Single-probe Variance Reduced Estimator for Coupled
Compositional Optimization [49.58290066287418]
We propose a novel method named Multi-block-probe Variance Reduced (MSVR) to alleviate the complexity of compositional problems.
Our results improve upon prior ones in several aspects, including the order of sample complexities and dependence on strongity.
arXiv Detail & Related papers (2022-07-18T12:03:26Z) - Paraformer: Fast and Accurate Parallel Transformer for
Non-autoregressive End-to-End Speech Recognition [62.83832841523525]
We propose a fast and accurate parallel transformer, termed Paraformer.
It accurately predicts the number of output tokens and extract hidden variables.
It can attain comparable performance to the state-of-the-art AR transformer, with more than 10x speedup.
arXiv Detail & Related papers (2022-06-16T17:24:14Z) - Asynchronous Stochastic Optimization Robust to Arbitrary Delays [54.61797739710608]
We consider optimization with delayed gradients where, at each time stept$, the algorithm makes an update using a stale computation - d_t$ for arbitrary delay $d_t gradient.
Our experiments demonstrate the efficacy and robustness of our algorithm in cases where the delay distribution is skewed or heavy-tailed.
arXiv Detail & Related papers (2021-06-22T15:50:45Z) - Faster Differentially Private Samplers via R\'enyi Divergence Analysis
of Discretized Langevin MCMC [35.050135428062795]
Langevin dynamics-based algorithms offer much faster alternatives under some distance measures such as statistical distance.
Our techniques simple and generic and apply to underdamped Langevin dynamics.
arXiv Detail & Related papers (2020-10-27T22:52:45Z) - Accelerated FBP for computed tomography image reconstruction [1.0266928164137636]
Filtered back projection (FBP) is a commonly used technique in tomographic image reconstruction demonstrating acceptable quality.
We propose a novel approach that reduces the computational complexity of the algorithm to $Theta(N2log N)$ addition operations avoiding Fourier space.
arXiv Detail & Related papers (2020-07-13T10:16:54Z) - DO-Conv: Depthwise Over-parameterized Convolutional Layer [66.46704754669169]
We propose to augment a convolutional layer with an additional depthwise convolution, where each input channel is convolved with a different 2D kernel.
We show with extensive experiments that the mere replacement of conventional convolutional layers with DO-Conv layers boosts the performance of CNNs.
arXiv Detail & Related papers (2020-06-22T06:57:10Z) - Adversarial Optimal Transport Through The Convolution Of Kernels With
Evolving Measures [3.1735221946062313]
A novel algorithm is proposed to solve the sample-based optimal transport problem.
The representation of the test function as the Monte Carlo simulation of a distribution makes the algorithm robust to dimensionality.
arXiv Detail & Related papers (2020-06-07T19:42:50Z) - Non-Adaptive Adaptive Sampling on Turnstile Streams [57.619901304728366]
We give the first relative-error algorithms for column subset selection, subspace approximation, projective clustering, and volume on turnstile streams that use space sublinear in $n$.
Our adaptive sampling procedure has a number of applications to various data summarization problems that either improve state-of-the-art or have only been previously studied in the more relaxed row-arrival model.
arXiv Detail & Related papers (2020-04-23T05:00:21Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.