IDF++: Analyzing and Improving Integer Discrete Flows for Lossless
Compression
- URL: http://arxiv.org/abs/2006.12459v2
- Date: Tue, 23 Mar 2021 09:40:50 GMT
- Title: IDF++: Analyzing and Improving Integer Discrete Flows for Lossless
Compression
- Authors: Rianne van den Berg, Alexey A. Gritsenko, Mostafa Dehghani, Casper
Kaae S{\o}nderby, Tim Salimans
- Abstract summary: discrete flows are a proposed class of models that learn invertible transformations for integer-valued random variables.
We show how different architecture modifications improve the performance of this model class for lossless compression.
- Score: 20.162897999101716
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In this paper we analyse and improve integer discrete flows for lossless
compression. Integer discrete flows are a recently proposed class of models
that learn invertible transformations for integer-valued random variables.
Their discrete nature makes them particularly suitable for lossless compression
with entropy coding schemes. We start by investigating a recent theoretical
claim that states that invertible flows for discrete random variables are less
flexible than their continuous counterparts. We demonstrate with a proof that
this claim does not hold for integer discrete flows due to the embedding of
data with finite support into the countably infinite integer lattice.
Furthermore, we zoom in on the effect of gradient bias due to the
straight-through estimator in integer discrete flows, and demonstrate that its
influence is highly dependent on architecture choices and less prominent than
previously thought. Finally, we show how different architecture modifications
improve the performance of this model class for lossless compression, and that
they also enable more efficient compression: a model with half the number of
flow layers performs on par with or better than the original integer discrete
flow model.
Related papers
- Unsupervised Representation Learning from Sparse Transformation Analysis [79.94858534887801]
We propose to learn representations from sequence data by factorizing the transformations of the latent variables into sparse components.
Input data are first encoded as distributions of latent activations and subsequently transformed using a probability flow model.
arXiv Detail & Related papers (2024-10-07T23:53:25Z) - Simple Cycle Reservoirs are Universal [0.358439716487063]
Reservoir models form a subclass of recurrent neural networks with fixed non-trainable input and dynamic coupling weights.
We show that they are capable of universal approximation of any unrestricted linear reservoir system.
arXiv Detail & Related papers (2023-08-21T15:35:59Z) - ChiroDiff: Modelling chirographic data with Diffusion Models [132.5223191478268]
We introduce a powerful model-class namely "Denoising Diffusion Probabilistic Models" or DDPMs for chirographic data.
Our model named "ChiroDiff", being non-autoregressive, learns to capture holistic concepts and therefore remains resilient to higher temporal sampling rate.
arXiv Detail & Related papers (2023-04-07T15:17:48Z) - ManiFlow: Implicitly Representing Manifolds with Normalizing Flows [145.9820993054072]
Normalizing Flows (NFs) are flexible explicit generative models that have been shown to accurately model complex real-world data distributions.
We propose an optimization objective that recovers the most likely point on the manifold given a sample from the perturbed distribution.
Finally, we focus on 3D point clouds for which we utilize the explicit nature of NFs, i.e. surface normals extracted from the gradient of the log-likelihood and the log-likelihood itself.
arXiv Detail & Related papers (2022-08-18T16:07:59Z) - Convolutional Persistence Transforms [0.6526824510982802]
We consider featurizations of data defined over simplicial complexes, like images and labeled graphs.
The persistence diagram of the resulting convolution describes the way the motif is distributed across the simplicial complex.
arXiv Detail & Related papers (2022-08-03T14:37:50Z) - Unified Multivariate Gaussian Mixture for Efficient Neural Image
Compression [151.3826781154146]
latent variables with priors and hyperpriors is an essential problem in variational image compression.
We find inter-correlations and intra-correlations exist when observing latent variables in a vectorized perspective.
Our model has better rate-distortion performance and an impressive $3.18times$ compression speed up.
arXiv Detail & Related papers (2022-03-21T11:44:17Z) - Discrete Denoising Flows [87.44537620217673]
We introduce a new discrete flow-based model for categorical random variables: Discrete Denoising Flows (DDFs)
In contrast with other discrete flow-based models, our model can be locally trained without introducing gradient bias.
We show that DDFs outperform Discrete Flows on modeling a toy example, binary MNIST and Cityscapes segmentation maps, measured in log-likelihood.
arXiv Detail & Related papers (2021-07-24T14:47:22Z) - iVPF: Numerical Invertible Volume Preserving Flow for Efficient Lossless
Compression [21.983560104199622]
It is nontrivial to store rapidly growing big data nowadays, which demands high-performance compression techniques.
We propose Numerical Invertible Volume Preserving Flow (iVPF) computation which is derived from the general volume preserving flows.
Experiments on various datasets show that the algorithm based on iVPF achieves state-of-the-art compression ratio over lightweight compression algorithms.
arXiv Detail & Related papers (2021-03-30T09:50:58Z) - Self Normalizing Flows [65.73510214694987]
We propose a flexible framework for training normalizing flows by replacing expensive terms in the gradient by learned approximate inverses at each layer.
This reduces the computational complexity of each layer's exact update from $mathcalO(D3)$ to $mathcalO(D2)$.
We show experimentally that such models are remarkably stable and optimize to similar data likelihood values as their exact gradient counterparts.
arXiv Detail & Related papers (2020-11-14T09:51:51Z) - Closing the Dequantization Gap: PixelCNN as a Single-Layer Flow [16.41460104376002]
We introduce subset flows, a class of flows that can transform finite volumes and allow exact computation of likelihoods for discrete data.
We identify ordinal discrete autoregressive models, including WaveNets, PixelCNNs and Transformers, as single-layer flows.
We demonstrate state-of-the-art results on CIFAR-10 for flow models trained with dequantization.
arXiv Detail & Related papers (2020-02-06T22:58:51Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.