Universal Rate-Distortion-Perception Representations for Lossy
Compression
- URL: http://arxiv.org/abs/2106.10311v1
- Date: Fri, 18 Jun 2021 18:52:08 GMT
- Title: Universal Rate-Distortion-Perception Representations for Lossy
Compression
- Authors: George Zhang, Jingjing Qian, Jun Chen, Ashish Khisti
- Abstract summary: We consider the notion of universal representations in which one may fix an encoder and vary the decoder to achieve any point within a collection of distortion and perception constraints.
We prove that the corresponding information-theoretic universal rate-distortion-perception is operationally achievable in an approximate sense.
- Score: 31.28856752892628
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: In the context of lossy compression, Blau & Michaeli (2019) adopt a
mathematical notion of perceptual quality and define the information
rate-distortion-perception function, generalizing the classical rate-distortion
tradeoff. We consider the notion of universal representations in which one may
fix an encoder and vary the decoder to achieve any point within a collection of
distortion and perception constraints. We prove that the corresponding
information-theoretic universal rate-distortion-perception function is
operationally achievable in an approximate sense. Under MSE distortion, we show
that the entire distortion-perception tradeoff of a Gaussian source can be
achieved by a single encoder of the same rate asymptotically. We then
characterize the achievable distortion-perception region for a fixed
representation in the case of arbitrary distributions, identify conditions
under which the aforementioned results continue to hold approximately, and
study the case when the rate is not fixed in advance. This motivates the study
of practical constructions that are approximately universal across the RDP
tradeoff, thereby alleviating the need to design a new encoder for each
objective. We provide experimental results on MNIST and SVHN suggesting that on
image compression tasks, the operational tradeoffs achieved by machine learning
models with a fixed encoder suffer only a small penalty when compared to their
variable encoder counterparts.
Related papers
- Sample what you cant compress [6.24979299238534]
We show how to learn a continuous encoder and decoder under a diffusion-based loss.
This approach yields better reconstruction quality as compared to GAN-based autoencoders.
We also show that the resulting representation is easier to model with a latent diffusion model as compared to the representation obtained from a state-of-the-art GAN-based loss.
arXiv Detail & Related papers (2024-09-04T08:42:42Z) - Once-for-All: Controllable Generative Image Compression with Dynamic Granularity Adaption [57.056311855630916]
We propose a Controllable Generative Image Compression framework, Control-GIC.
It is capable of fine-grained adaption across a broad spectrum while ensuring high-fidelity and generality compression.
We develop a conditional conditionalization that can trace back to historic encoded multi-granularity representations.
arXiv Detail & Related papers (2024-06-02T14:22:09Z) - Correcting Diffusion-Based Perceptual Image Compression with Privileged End-to-End Decoder [49.01721042973929]
This paper presents a diffusion-based image compression method that employs a privileged end-to-end decoder model as correction.
Experiments demonstrate the superiority of our method in both distortion and perception compared with previous perceptual compression methods.
arXiv Detail & Related papers (2024-04-07T10:57:54Z) - Optimally Controllable Perceptual Lossy Compression [16.208548355509127]
Recent studies in lossy compression show that distortion and perceptual quality are at odds with each other.
To attain different perceptual quality, different decoders have to be trained.
We present a nontrivial finding that only two decoders are sufficient for optimally achieving arbitrary D-P tradeoffs.
arXiv Detail & Related papers (2022-06-21T02:48:35Z) - Neural Estimation of the Rate-Distortion Function With Applications to
Operational Source Coding [25.59334941818991]
A fundamental question in designing lossy data compression schemes is how well one can do in comparison with the rate-distortion function.
We investigate methods to estimate the rate-distortion function on large, real-world data.
We apply the resulting rate-distortion estimator, called NERD, on popular image datasets.
arXiv Detail & Related papers (2022-04-04T16:06:40Z) - Reducing Redundancy in the Bottleneck Representation of the Autoencoders [98.78384185493624]
Autoencoders are a type of unsupervised neural networks, which can be used to solve various tasks.
We propose a scheme to explicitly penalize feature redundancies in the bottleneck representation.
We tested our approach across different tasks: dimensionality reduction using three different dataset, image compression using the MNIST dataset, and image denoising using fashion MNIST.
arXiv Detail & Related papers (2022-02-09T18:48:02Z) - On Perceptual Lossy Compression: The Cost of Perceptual Reconstruction
and An Optimal Training Framework [12.13586501618741]
We show that the cost of achieving perfect perception quality is exactly a doubling of the lowest achievable MSE distortion.
We propose a novel training framework to achieve the lowest MSE distortion under perfect perception constraint at a given bit rate.
arXiv Detail & Related papers (2021-06-05T02:53:38Z) - Substitutional Neural Image Compression [48.20906717052056]
Substitutional Neural Image Compression (SNIC) is a general approach for enhancing any neural image compression model.
It boosts compression performance toward a flexible distortion metric and enables bit-rate control using a single model instance.
arXiv Detail & Related papers (2021-05-16T20:53:31Z) - Autoencoding Variational Autoencoder [56.05008520271406]
We study the implications of this behaviour on the learned representations and also the consequences of fixing it by introducing a notion of self consistency.
We show that encoders trained with our self-consistency approach lead to representations that are robust (insensitive) to perturbations in the input introduced by adversarial attacks.
arXiv Detail & Related papers (2020-12-07T14:16:14Z) - The Rate-Distortion-Accuracy Tradeoff: JPEG Case Study [30.84385779593074]
We focus on the design of the quantization tables in the JPEG compression standard.
We offer a novel optimal tuning of these tables via continuous optimization.
We report a substantial boost in performance by a simple and easily implemented modification of these tables.
arXiv Detail & Related papers (2020-08-03T01:39:01Z) - A Deep Ordinal Distortion Estimation Approach for Distortion Rectification [62.72089758481803]
We propose a novel distortion rectification approach that can obtain more accurate parameters with higher efficiency.
We design a local-global associated estimation network that learns the ordinal distortion to approximate the realistic distortion distribution.
Considering the redundancy of distortion information, our approach only uses a part of distorted image for the ordinal distortion estimation.
arXiv Detail & Related papers (2020-07-21T10:03:42Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.