Variable-Rate Learned Image Compression with Multi-Objective
Optimization and Quantization-Reconstruction Offsets
- URL: http://arxiv.org/abs/2402.18930v1
- Date: Thu, 29 Feb 2024 07:45:02 GMT
- Title: Variable-Rate Learned Image Compression with Multi-Objective
Optimization and Quantization-Reconstruction Offsets
- Authors: Fatih Kamisli, Fabien Racape, Hyomin Choi
- Abstract summary: This paper follows the traditional approach to vary a single quantization step size to perform uniform quantization of all latent tensor elements.
Three modifications are proposed to improve the variable rate compression performance.
The achieved variable rate compression results indicate negligible or minimal compression performance loss compared to training multiple models.
- Score: 8.670873561640903
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Achieving successful variable bitrate compression with computationally simple
algorithms from a single end-to-end learned image or video compression model
remains a challenge. Many approaches have been proposed, including conditional
auto-encoders, channel-adaptive gains for the latent tensor or uniformly
quantizing all elements of the latent tensor. This paper follows the
traditional approach to vary a single quantization step size to perform uniform
quantization of all latent tensor elements. However, three modifications are
proposed to improve the variable rate compression performance. First, multi
objective optimization is used for (post) training. Second, a
quantization-reconstruction offset is introduced into the quantization
operation. Third, variable rate quantization is used also for the hyper latent.
All these modifications can be made on a pre-trained single-rate compression
model by performing post training. The algorithms are implemented into three
well-known image compression models and the achieved variable rate compression
results indicate negligible or minimal compression performance loss compared to
training multiple models. (Codes will be shared at
https://github.com/InterDigitalInc/CompressAI)
Related papers
- DiffRate : Differentiable Compression Rate for Efficient Vision
Transformers [98.33906104846386]
Token compression aims to speed up large-scale vision transformers (e.g. ViTs) by pruning (dropping) or merging tokens.
DiffRate is a novel token compression method that has several appealing properties prior arts do not have.
arXiv Detail & Related papers (2023-05-29T10:15:19Z) - Multiscale Augmented Normalizing Flows for Image Compression [17.441496966834933]
We present a novel concept, which adapts the hierarchical latent space for augmented normalizing flows, an invertible latent variable model.
Our best performing model achieved average rate savings of more than 7% over comparable single-scale models.
arXiv Detail & Related papers (2023-05-09T13:42:43Z) - Selective compression learning of latent representations for
variable-rate image compression [38.077284943341105]
We propose a selective compression method that partially encodes latent representations in a fully generalized manner for deep learning-based variable-rate image compression.
The proposed method can achieve comparable compression efficiency as those of the separately trained reference compression models and can reduce decoding time owing to the selective compression.
arXiv Detail & Related papers (2022-11-08T09:09:59Z) - High-Fidelity Variable-Rate Image Compression via Invertible Activation
Transformation [24.379052026260034]
We propose the Invertible Activation Transformation (IAT) module to tackle the issue of high-fidelity fine variable-rate image compression.
IAT and QLevel together give the image compression model the ability of fine variable-rate control while better maintaining the image fidelity.
Our method outperforms the state-of-the-art variable-rate image compression method by a large margin, especially after multiple re-encodings.
arXiv Detail & Related papers (2022-09-12T07:14:07Z) - OPQ: Compressing Deep Neural Networks with One-shot Pruning-Quantization [32.60139548889592]
We propose a novel One-shot Pruning-Quantization (OPQ) in this paper.
OPQ analytically solves the compression allocation with pre-trained weight parameters only.
We propose a unified channel-wise quantization method that enforces all channels of each layer to share a common codebook.
arXiv Detail & Related papers (2022-05-23T09:05:25Z) - Unified Multivariate Gaussian Mixture for Efficient Neural Image
Compression [151.3826781154146]
latent variables with priors and hyperpriors is an essential problem in variational image compression.
We find inter-correlations and intra-correlations exist when observing latent variables in a vectorized perspective.
Our model has better rate-distortion performance and an impressive $3.18times$ compression speed up.
arXiv Detail & Related papers (2022-03-21T11:44:17Z) - Variable-Rate Deep Image Compression through Spatially-Adaptive Feature
Transform [58.60004238261117]
We propose a versatile deep image compression network based on Spatial Feature Transform (SFT arXiv:1804.02815)
Our model covers a wide range of compression rates using a single model, which is controlled by arbitrary pixel-wise quality maps.
The proposed framework allows us to perform task-aware image compressions for various tasks.
arXiv Detail & Related papers (2021-08-21T17:30:06Z) - Differentiable Model Compression via Pseudo Quantization Noise [99.89011673907814]
We propose to add independent pseudo quantization noise to model parameters during training to approximate the effect of a quantization operator.
We experimentally verify that our method outperforms state-of-the-art quantization techniques on several benchmarks and architectures for image classification, language modeling, and audio source separation.
arXiv Detail & Related papers (2021-04-20T14:14:03Z) - Permute, Quantize, and Fine-tune: Efficient Compression of Neural
Networks [70.0243910593064]
Key to success of vector quantization is deciding which parameter groups should be compressed together.
In this paper we make the observation that the weights of two adjacent layers can be permuted while expressing the same function.
We then establish a connection to rate-distortion theory and search for permutations that result in networks that are easier to compress.
arXiv Detail & Related papers (2020-10-29T15:47:26Z) - Training with Quantization Noise for Extreme Model Compression [57.51832088938618]
We tackle the problem of producing compact models, maximizing their accuracy for a given model size.
A standard solution is to train networks with Quantization Aware Training, where the weights are quantized during training and the gradients approximated with the Straight-Through Estimator.
In this paper, we extend this approach to work beyond int8 fixed-point quantization with extreme compression methods.
arXiv Detail & Related papers (2020-04-15T20:10:53Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.