Exploring Resolution Fields for Scalable Image Compression with
Uncertainty Guidance
- URL: http://arxiv.org/abs/2306.08941v1
- Date: Thu, 15 Jun 2023 08:26:24 GMT
- Title: Exploring Resolution Fields for Scalable Image Compression with
Uncertainty Guidance
- Authors: Dongyi Zhang, Feng Li, Man Liu, Runmin Cong, Huihui Bai, Meng Wang and
Yao Zhao
- Abstract summary: In this work, we explore the potential of resolution fields in scalable image compression.
We propose the reciprocal pyramid network (RPN) that fulfills the need for more adaptable and versatile compression.
Experiments show the superiority of RPN against existing classical and deep learning-based scalable codecs.
- Score: 47.96024424475888
- License: http://arxiv.org/licenses/nonexclusive-distrib/1.0/
- Abstract: Recently, there are significant advancements in learning-based image
compression methods surpassing traditional coding standards. Most of them
prioritize achieving the best rate-distortion performance for a particular
compression rate, which limits their flexibility and adaptability in various
applications with complex and varying constraints. In this work, we explore the
potential of resolution fields in scalable image compression and propose the
reciprocal pyramid network (RPN) that fulfills the need for more adaptable and
versatile compression. Specifically, RPN first builds a compression pyramid and
generates the resolution fields at different levels in a top-down manner. The
key design lies in the cross-resolution context mining module between adjacent
levels, which performs feature enriching and distillation to mine meaningful
contextualized information and remove unnecessary redundancy, producing
informative resolution fields as residual priors. The scalability is achieved
by progressive bitstream reusing and resolution field incorporation varying at
different levels. Furthermore, between adjacent compression levels, we
explicitly quantify the aleatoric uncertainty from the bottom decoded
representations and develop an uncertainty-guided loss to update the
upper-level compression parameters, forming a reverse pyramid process that
enforces the network to focus on the textured pixels with high variance for
more reliable and accurate reconstruction. Combining resolution field
exploration and uncertainty guidance in a pyramid manner, RPN can effectively
achieve spatial and quality scalable image compression. Experiments show the
superiority of RPN against existing classical and deep learning-based scalable
codecs. Code will be available at https://github.com/JGIroro/RPNSIC.
Related papers
- Exploiting Inter-Image Similarity Prior for Low-Bitrate Remote Sensing Image Compression [10.427300958330816]
We propose a codebook-based RS image compression (Code-RSIC) method with a generated discrete codebook.
The code significantly outperforms state-of-the-art traditional and learning-based image compression algorithms in terms of perception quality.
arXiv Detail & Related papers (2024-07-17T03:33:16Z) - Once-for-All: Controllable Generative Image Compression with Dynamic Granularity Adaption [57.056311855630916]
We propose a Controllable Generative Image Compression framework, Control-GIC.
It is capable of fine-grained adaption across a broad spectrum while ensuring high-fidelity and generality compression.
We develop a conditional conditionalization that can trace back to historic encoded multi-granularity representations.
arXiv Detail & Related papers (2024-06-02T14:22:09Z) - Extreme Image Compression using Fine-tuned VQGANs [43.43014096929809]
We introduce vector quantization (VQ)-based generative models into the image compression domain.
The codebook learned by the VQGAN model yields a strong expressive capacity.
The proposed framework outperforms state-of-the-art codecs in terms of perceptual quality-oriented metrics.
arXiv Detail & Related papers (2023-07-17T06:14:19Z) - Beyond Learned Metadata-based Raw Image Reconstruction [86.1667769209103]
Raw images have distinct advantages over sRGB images, e.g., linearity and fine-grained quantization levels.
They are not widely adopted by general users due to their substantial storage requirements.
We propose a novel framework that learns a compact representation in the latent space, serving as metadata.
arXiv Detail & Related papers (2023-06-21T06:59:07Z) - DeepFGS: Fine-Grained Scalable Coding for Learned Image Compression [22.933872281183497]
We propose the first learned fine-grained scalable image compression model (DeepFGS)
In this paper, we introduce a feature separation backbone to divide the image information into basic and scalable features, then redistribute the features channel by channel through an information rearrangement strategy.
Experiments demonstrate that our DeepFGS outperforms all learning-based scalable image compression models and conventional scalable image codecs in PSNR and MS-SSIM metrics.
arXiv Detail & Related papers (2022-01-04T15:03:13Z) - Implicit Neural Representations for Image Compression [103.78615661013623]
Implicit Neural Representations (INRs) have gained attention as a novel and effective representation for various data types.
We propose the first comprehensive compression pipeline based on INRs including quantization, quantization-aware retraining and entropy coding.
We find that our approach to source compression with INRs vastly outperforms similar prior work.
arXiv Detail & Related papers (2021-12-08T13:02:53Z) - Variable-Rate Deep Image Compression through Spatially-Adaptive Feature
Transform [58.60004238261117]
We propose a versatile deep image compression network based on Spatial Feature Transform (SFT arXiv:1804.02815)
Our model covers a wide range of compression rates using a single model, which is controlled by arbitrary pixel-wise quality maps.
The proposed framework allows us to perform task-aware image compressions for various tasks.
arXiv Detail & Related papers (2021-08-21T17:30:06Z) - Enhanced Invertible Encoding for Learned Image Compression [40.21904131503064]
In this paper, we propose an enhanced Invertible.
Network with invertible neural networks (INNs) to largely mitigate the information loss problem for better compression.
Experimental results on the Kodak, CLIC, and Tecnick datasets show that our method outperforms the existing learned image compression methods.
arXiv Detail & Related papers (2021-08-08T17:32:10Z) - Modeling Lost Information in Lossy Image Compression [72.69327382643549]
Lossy image compression is one of the most commonly used operators for digital images.
We propose a novel invertible framework called Invertible Lossy Compression (ILC) to largely mitigate the information loss problem.
arXiv Detail & Related papers (2020-06-22T04:04:56Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.