Approaching Rate-Distortion Limits in Neural Compression with Lattice
Transform Coding
- URL: http://arxiv.org/abs/2403.07320v1
- Date: Tue, 12 Mar 2024 05:09:25 GMT
- Title: Approaching Rate-Distortion Limits in Neural Compression with Lattice
Transform Coding
- Authors: Eric Lei, Hamed Hassani, Shirin Saeedi Bidokhti
- Abstract summary: neural compression design involves transforming the source to a latent vector, which is then rounded to integers and entropy coded.
We show that it is highly sub-optimal on i.i.d. sequences, and in fact always recovers scalar quantization of the original source sequence.
By employing lattice quantization instead of scalar quantization in the latent space, we demonstrate that Lattice Transform Coding (LTC) is able to recover optimal vector quantization at various dimensions.
- Score: 33.377272636443344
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Neural compression has brought tremendous progress in designing lossy
compressors with good rate-distortion (RD) performance at low complexity. Thus
far, neural compression design involves transforming the source to a latent
vector, which is then rounded to integers and entropy coded. While this
approach has been shown to be optimal in a one-shot sense on certain sources,
we show that it is highly sub-optimal on i.i.d. sequences, and in fact always
recovers scalar quantization of the original source sequence. We demonstrate
that the sub-optimality is due to the choice of quantization scheme in the
latent space, and not the transform design. By employing lattice quantization
instead of scalar quantization in the latent space, we demonstrate that Lattice
Transform Coding (LTC) is able to recover optimal vector quantization at
various dimensions and approach the asymptotically-achievable rate-distortion
function at reasonable complexity. On general vector sources, LTC improves upon
standard neural compressors in one-shot coding performance. LTC also enables
neural compressors that perform block coding on i.i.d. vector sources, which
yields coding gain over optimal one-shot coding.
Related papers
- Frequency Disentangled Features in Neural Image Compression [13.016298207860974]
A neural image compression network is governed by how well the entropy model matches the true distribution of the latent code.
In this paper, we propose a feature-level frequency disentanglement to help the relaxed scalar quantization achieve lower bit rates.
The proposed network not only outperforms hand-engineered codecs, but also neural network-based codecs built on-heavy spatially autoregressive entropy models.
arXiv Detail & Related papers (2023-08-04T14:55:44Z) - Joint Hierarchical Priors and Adaptive Spatial Resolution for Efficient
Neural Image Compression [11.25130799452367]
We propose an absolute image compression transformer (ICT) for neural image compression (NIC)
ICT captures both global and local contexts from the latent representations and better parameterize the distribution of the quantized latents.
Our framework significantly improves the trade-off between coding efficiency and decoder complexity over the versatile video coding (VVC) reference encoder (VTM-18.0) and the neural SwinT-ChARM.
arXiv Detail & Related papers (2023-07-05T13:17:14Z) - NVTC: Nonlinear Vector Transform Coding [35.10187626615328]
In theory, vector quantization (VQ) is always better than scalar quantization (SQ) in terms of rate-distortion (R-D) performance.
Recent state-of-the-art methods for neural image compression are mainly based on nonlinear transform coding (NTC) with uniform scalar quantization.
We propose a novel framework for neural image compression named Vector Transform Coding (NVTC)
arXiv Detail & Related papers (2023-05-25T13:06:38Z) - LVQAC: Lattice Vector Quantization Coupled with Spatially Adaptive
Companding for Efficient Learned Image Compression [24.812267280543693]
We present a novel Lattice Vector Quantization scheme coupled with a spatially Adaptive Companding (LVQAC) mapping.
For any end-to-end CNN image compression models, replacing uniform quantizer by LVQAC achieves better rate-distortion performance without significantly increasing the model complexity.
arXiv Detail & Related papers (2023-03-25T23:34:15Z) - Deep Quantum Error Correction [73.54643419792453]
Quantum error correction codes (QECC) are a key component for realizing the potential of quantum computing.
In this work, we efficiently train novel emphend-to-end deep quantum error decoders.
The proposed method demonstrates the power of neural decoders for QECC by achieving state-of-the-art accuracy.
arXiv Detail & Related papers (2023-01-27T08:16:26Z) - Unified Multivariate Gaussian Mixture for Efficient Neural Image
Compression [151.3826781154146]
latent variables with priors and hyperpriors is an essential problem in variational image compression.
We find inter-correlations and intra-correlations exist when observing latent variables in a vectorized perspective.
Our model has better rate-distortion performance and an impressive $3.18times$ compression speed up.
arXiv Detail & Related papers (2022-03-21T11:44:17Z) - Neural JPEG: End-to-End Image Compression Leveraging a Standard JPEG
Encoder-Decoder [73.48927855855219]
We propose a system that learns to improve the encoding performance by enhancing its internal neural representations on both the encoder and decoder ends.
Experiments demonstrate that our approach successfully improves the rate-distortion performance over JPEG across various quality metrics.
arXiv Detail & Related papers (2022-01-27T20:20:03Z) - Implicit Neural Representations for Image Compression [103.78615661013623]
Implicit Neural Representations (INRs) have gained attention as a novel and effective representation for various data types.
We propose the first comprehensive compression pipeline based on INRs including quantization, quantization-aware retraining and entropy coding.
We find that our approach to source compression with INRs vastly outperforms similar prior work.
arXiv Detail & Related papers (2021-12-08T13:02:53Z) - Learned transform compression with optimized entropy encoding [72.20409648915398]
We consider the problem of learned transform compression where we learn both, the transform and the probability distribution over the discrete codes.
We employ a soft relaxation of the quantization operation to allow for back-propagation of gradients and employ vector (rather than scalar) quantization of the latent codes.
arXiv Detail & Related papers (2021-04-07T17:58:01Z) - Unfolding Neural Networks for Compressive Multichannel Blind
Deconvolution [71.29848468762789]
We propose a learned-structured unfolding neural network for the problem of compressive sparse multichannel blind-deconvolution.
In this problem, each channel's measurements are given as convolution of a common source signal and sparse filter.
We demonstrate that our method is superior to classical structured compressive sparse multichannel blind-deconvolution methods in terms of accuracy and speed of sparse filter recovery.
arXiv Detail & Related papers (2020-10-22T02:34:33Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.