High-Fidelity Generative Image Compression
- URL: http://arxiv.org/abs/2006.09965v3
- Date: Fri, 23 Oct 2020 08:55:23 GMT
- Title: High-Fidelity Generative Image Compression
- Authors: Fabian Mentzer, George Toderici, Michael Tschannen, Eirikur Agustsson
- Abstract summary: We study how to combine Gene Adrial Networks and learned compression to obtain a state-of-the-art generative lossy compression system.
In particular, we investigate normalization layers, generator and discriminator architectures, training strategies, as well as perceptual losses.
- Score: 39.04379573099481
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: We extensively study how to combine Generative Adversarial Networks and
learned compression to obtain a state-of-the-art generative lossy compression
system. In particular, we investigate normalization layers, generator and
discriminator architectures, training strategies, as well as perceptual losses.
In contrast to previous work, i) we obtain visually pleasing reconstructions
that are perceptually similar to the input, ii) we operate in a broad range of
bitrates, and iii) our approach can be applied to high-resolution images. We
bridge the gap between rate-distortion-perception theory and practice by
evaluating our approach both quantitatively with various perceptual metrics,
and with a user study. The study shows that our method is preferred to previous
approaches even if they use more than 2x the bitrate.
Related papers
- Rate-Distortion-Cognition Controllable Versatile Neural Image Compression [47.72668401825835]
We propose a rate-distortion-cognition controllable versatile image compression method.
Our method yields satisfactory ICM performance and flexible Rate-DistortionCognition controlling.
arXiv Detail & Related papers (2024-07-16T13:17:51Z) - Semantic Ensemble Loss and Latent Refinement for High-Fidelity Neural Image Compression [58.618625678054826]
This study presents an enhanced neural compression method designed for optimal visual fidelity.
We have trained our model with a sophisticated semantic ensemble loss, integrating Charbonnier loss, perceptual loss, style loss, and a non-binary adversarial loss.
Our empirical findings demonstrate that this approach significantly improves the statistical fidelity of neural image compression.
arXiv Detail & Related papers (2024-01-25T08:11:27Z) - Transferable Learned Image Compression-Resistant Adversarial Perturbations [66.46470251521947]
Adversarial attacks can readily disrupt the image classification system, revealing the vulnerability of DNN-based recognition tasks.
We introduce a new pipeline that targets image classification models that utilize learned image compressors as pre-processing modules.
arXiv Detail & Related papers (2024-01-06T03:03:28Z) - Machine Perception-Driven Image Compression: A Layered Generative
Approach [32.23554195427311]
layered generative image compression model is proposed to achieve high human vision-oriented image reconstructed quality.
Task-agnostic learning-based compression model is proposed, which effectively supports various compressed domain-based analytical tasks.
Joint optimization schedule is adopted to acquire best balance point among compression ratio, reconstructed image quality, and downstream perception performance.
arXiv Detail & Related papers (2023-04-14T02:12:38Z) - Compressive Ptychography using Deep Image and Generative Priors [9.658250977094562]
Ptychography is a well-established coherent diffraction imaging technique that enables non-invasive imaging of samples at a nanometer scale.
One major limitation of ptychography is the long data acquisition time due to mechanical scanning of the sample.
We propose a generative model combining deep image priors with deep generative priors.
arXiv Detail & Related papers (2022-05-05T02:18:26Z) - Recognition-Aware Learned Image Compression [0.5801044612920815]
We propose a recognition-aware learned compression method, which optimize a rate-distortion loss alongside a task-specific loss.
Our method achieves 26% higher recognition accuracy at equivalents compared to traditional methods such as BPG.
arXiv Detail & Related papers (2022-02-01T03:33:51Z) - Implicit Neural Representations for Image Compression [103.78615661013623]
Implicit Neural Representations (INRs) have gained attention as a novel and effective representation for various data types.
We propose the first comprehensive compression pipeline based on INRs including quantization, quantization-aware retraining and entropy coding.
We find that our approach to source compression with INRs vastly outperforms similar prior work.
arXiv Detail & Related papers (2021-12-08T13:02:53Z) - Deblurring via Stochastic Refinement [85.42730934561101]
We present an alternative framework for blind deblurring based on conditional diffusion models.
Our method is competitive in terms of distortion metrics such as PSNR.
arXiv Detail & Related papers (2021-12-05T04:36:09Z) - Improving Inference for Neural Image Compression [31.999462074510305]
State-of-the-art methods build on hierarchical variational autoencoders to predict a compressible latent representation of each data point.
We identify three approximation gaps which limit performance in the conventional approach.
We propose remedies for each of these three limitations based on ideas related to iterative inference.
arXiv Detail & Related papers (2020-06-07T19:26:37Z) - Learning End-to-End Lossy Image Compression: A Benchmark [90.35363142246806]
We first conduct a comprehensive literature survey of learned image compression methods.
We describe milestones in cutting-edge learned image-compression methods, review a broad range of existing works, and provide insights into their historical development routes.
By introducing a coarse-to-fine hyperprior model for entropy estimation and signal reconstruction, we achieve improved rate-distortion performance.
arXiv Detail & Related papers (2020-02-10T13:13:43Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.