Split Hierarchical Variational Compression
- URL: http://arxiv.org/abs/2204.02071v1
- Date: Tue, 5 Apr 2022 09:13:38 GMT
- Title: Split Hierarchical Variational Compression
- Authors: Tom Ryder, Chen Zhang, Ning Kang, Shifeng Zhang
- Abstract summary: Variational autoencoders (VAEs) have witnessed great success in performing the compression of image datasets.
SHVC introduces an efficient autoregressive sub-pixel convolution, that allows a generalisation between per-pixel autoregressions and fully factorised probability models.
- Score: 21.474095984110622
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Variational autoencoders (VAEs) have witnessed great success in performing
the compression of image datasets. This success, made possible by the bits-back
coding framework, has produced competitive compression performance across many
benchmarks. However, despite this, VAE architectures are currently limited by a
combination of coding practicalities and compression ratios. That is, not only
do state-of-the-art methods, such as normalizing flows, often demonstrate
out-performance, but the initial bits required in coding makes single and
parallel image compression challenging. To remedy this, we introduce Split
Hierarchical Variational Compression (SHVC). SHVC introduces two novelties.
Firstly, we propose an efficient autoregressive prior, the autoregressive
sub-pixel convolution, that allows a generalisation between per-pixel
autoregressions and fully factorised probability models. Secondly, we define
our coding framework, the autoregressive initial bits, that flexibly supports
parallel coding and avoids -- for the first time -- many of the practicalities
commonly associated with bits-back coding. In our experiments, we demonstrate
SHVC is able to achieve state-of-the-art compression performance across
full-resolution lossless image compression tasks, with up to 100x fewer model
parameters than competing VAE approaches.
Related papers
- Once-for-All: Controllable Generative Image Compression with Dynamic Granularity Adaption [57.056311855630916]
We propose a Controllable Generative Image Compression framework, Control-GIC.
It is capable of fine-grained adaption across a broad spectrum while ensuring high-fidelity and generality compression.
We develop a conditional conditionalization that can trace back to historic encoded multi-granularity representations.
arXiv Detail & Related papers (2024-06-02T14:22:09Z) - Extreme Image Compression using Fine-tuned VQGANs [43.43014096929809]
We introduce vector quantization (VQ)-based generative models into the image compression domain.
The codebook learned by the VQGAN model yields a strong expressive capacity.
The proposed framework outperforms state-of-the-art codecs in terms of perceptual quality-oriented metrics.
arXiv Detail & Related papers (2023-07-17T06:14:19Z) - You Can Mask More For Extremely Low-Bitrate Image Compression [80.7692466922499]
Learned image compression (LIC) methods have experienced significant progress during recent years.
LIC methods fail to explicitly explore the image structure and texture components crucial for image compression.
We present DA-Mask that samples visible patches based on the structure and texture of original images.
We propose a simple yet effective masked compression model (MCM), the first framework that unifies LIC and LIC end-to-end for extremely low-bitrate compression.
arXiv Detail & Related papers (2023-06-27T15:36:22Z) - Multiscale Augmented Normalizing Flows for Image Compression [17.441496966834933]
We present a novel concept, which adapts the hierarchical latent space for augmented normalizing flows, an invertible latent variable model.
Our best performing model achieved average rate savings of more than 7% over comparable single-scale models.
arXiv Detail & Related papers (2023-05-09T13:42:43Z) - Device Interoperability for Learned Image Compression with Weights and
Activations Quantization [1.373801677008598]
We present a method to solve the device interoperability problem of a state-of-the-art image compression network.
We suggest a simple method which can ensure cross-platform encoding and decoding, and can be implemented quickly.
arXiv Detail & Related papers (2022-12-02T17:45:29Z) - Deep Lossy Plus Residual Coding for Lossless and Near-lossless Image
Compression [85.93207826513192]
We propose a unified and powerful deep lossy plus residual (DLPR) coding framework for both lossless and near-lossless image compression.
We solve the joint lossy and residual compression problem in the approach of VAEs.
In the near-lossless mode, we quantize the original residuals to satisfy a given $ell_infty$ error bound.
arXiv Detail & Related papers (2022-09-11T12:11:56Z) - ELIC: Efficient Learned Image Compression with Unevenly Grouped
Space-Channel Contextual Adaptive Coding [9.908820641439368]
We propose an efficient model, ELIC, to achieve state-of-the-art speed and compression ability.
With superior performance, the proposed model also supports extremely fast preview decoding and progressive decoding.
arXiv Detail & Related papers (2022-03-21T11:19:50Z) - Implicit Neural Representations for Image Compression [103.78615661013623]
Implicit Neural Representations (INRs) have gained attention as a novel and effective representation for various data types.
We propose the first comprehensive compression pipeline based on INRs including quantization, quantization-aware retraining and entropy coding.
We find that our approach to source compression with INRs vastly outperforms similar prior work.
arXiv Detail & Related papers (2021-12-08T13:02:53Z) - Lossless Compression with Latent Variable Models [4.289574109162585]
We use latent variable models, which we call 'bits back with asymmetric numeral systems' (BB-ANS)
The method involves interleaving encode and decode steps, and achieves an optimal rate when compressing batches of data.
We describe 'Craystack', a modular software framework which we have developed for rapid prototyping of compression using deep generative models.
arXiv Detail & Related papers (2021-04-21T14:03:05Z) - Conditional Entropy Coding for Efficient Video Compression [82.35389813794372]
We propose a very simple and efficient video compression framework that only focuses on modeling the conditional entropy between frames.
We first show that a simple architecture modeling the entropy between the image latent codes is as competitive as other neural video compression works and video codecs.
We then propose a novel internal learning extension on top of this architecture that brings an additional 10% savings without trading off decoding speed.
arXiv Detail & Related papers (2020-08-20T20:01:59Z) - Learning End-to-End Lossy Image Compression: A Benchmark [90.35363142246806]
We first conduct a comprehensive literature survey of learned image compression methods.
We describe milestones in cutting-edge learned image-compression methods, review a broad range of existing works, and provide insights into their historical development routes.
By introducing a coarse-to-fine hyperprior model for entropy estimation and signal reconstruction, we achieve improved rate-distortion performance.
arXiv Detail & Related papers (2020-02-10T13:13:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.