Variable-Rate Deep Image Compression through Spatially-Adaptive Feature
Transform
- URL: http://arxiv.org/abs/2108.09551v1
- Date: Sat, 21 Aug 2021 17:30:06 GMT
- Title: Variable-Rate Deep Image Compression through Spatially-Adaptive Feature
Transform
- Authors: Myungseo Song, Jinyoung Choi, Bohyung Han
- Abstract summary: We propose a versatile deep image compression network based on Spatial Feature Transform (SFT arXiv:1804.02815)
Our model covers a wide range of compression rates using a single model, which is controlled by arbitrary pixel-wise quality maps.
The proposed framework allows us to perform task-aware image compressions for various tasks.
- Score: 58.60004238261117
- License: http://creativecommons.org/licenses/by-nc-sa/4.0/
- Abstract: We propose a versatile deep image compression network based on Spatial
Feature Transform (SFT arXiv:1804.02815), which takes a source image and a
corresponding quality map as inputs and produce a compressed image with
variable rates. Our model covers a wide range of compression rates using a
single model, which is controlled by arbitrary pixel-wise quality maps. In
addition, the proposed framework allows us to perform task-aware image
compressions for various tasks, e.g., classification, by efficiently estimating
optimized quality maps specific to target tasks for our encoding network. This
is even possible with a pretrained network without learning separate models for
individual tasks. Our algorithm achieves outstanding rate-distortion trade-off
compared to the approaches based on multiple models that are optimized
separately for several different target rates. At the same level of
compression, the proposed approach successfully improves performance on image
classification and text region quality preservation via task-aware quality map
estimation without additional model training. The code is available at the
project website: https://github.com/micmic123/QmapCompression
Related papers
- Progressive Learning with Visual Prompt Tuning for Variable-Rate Image
Compression [60.689646881479064]
We propose a progressive learning paradigm for transformer-based variable-rate image compression.
Inspired by visual prompt tuning, we use LPM to extract prompts for input images and hidden features at the encoder side and decoder side, respectively.
Our model outperforms all current variable image methods in terms of rate-distortion performance and approaches the state-of-the-art fixed image compression methods trained from scratch.
arXiv Detail & Related papers (2023-11-23T08:29:32Z) - Beyond Learned Metadata-based Raw Image Reconstruction [86.1667769209103]
Raw images have distinct advantages over sRGB images, e.g., linearity and fine-grained quantization levels.
They are not widely adopted by general users due to their substantial storage requirements.
We propose a novel framework that learns a compact representation in the latent space, serving as metadata.
arXiv Detail & Related papers (2023-06-21T06:59:07Z) - Multiscale Augmented Normalizing Flows for Image Compression [17.441496966834933]
We present a novel concept, which adapts the hierarchical latent space for augmented normalizing flows, an invertible latent variable model.
Our best performing model achieved average rate savings of more than 7% over comparable single-scale models.
arXiv Detail & Related papers (2023-05-09T13:42:43Z) - Selective compression learning of latent representations for
variable-rate image compression [38.077284943341105]
We propose a selective compression method that partially encodes latent representations in a fully generalized manner for deep learning-based variable-rate image compression.
The proposed method can achieve comparable compression efficiency as those of the separately trained reference compression models and can reduce decoding time owing to the selective compression.
arXiv Detail & Related papers (2022-11-08T09:09:59Z) - High-Fidelity Variable-Rate Image Compression via Invertible Activation
Transformation [24.379052026260034]
We propose the Invertible Activation Transformation (IAT) module to tackle the issue of high-fidelity fine variable-rate image compression.
IAT and QLevel together give the image compression model the ability of fine variable-rate control while better maintaining the image fidelity.
Our method outperforms the state-of-the-art variable-rate image compression method by a large margin, especially after multiple re-encodings.
arXiv Detail & Related papers (2022-09-12T07:14:07Z) - Estimating the Resize Parameter in End-to-end Learned Image Compression [50.20567320015102]
We describe a search-free resizing framework that can further improve the rate-distortion tradeoff of recent learned image compression models.
Our results show that our new resizing parameter estimation framework can provide Bjontegaard-Delta rate (BD-rate) improvement of about 10% against leading perceptual quality engines.
arXiv Detail & Related papers (2022-04-26T01:35:02Z) - Early Exit or Not: Resource-Efficient Blind Quality Enhancement for
Compressed Images [54.40852143927333]
Lossy image compression is pervasively conducted to save communication bandwidth, resulting in undesirable compression artifacts.
We propose a resource-efficient blind quality enhancement (RBQE) approach for compressed images.
Our approach can automatically decide to terminate or continue enhancement according to the assessed quality of enhanced images.
arXiv Detail & Related papers (2020-06-30T07:38:47Z) - Quantization Guided JPEG Artifact Correction [69.04777875711646]
We develop a novel architecture for artifact correction using the JPEG files quantization matrix.
This allows our single model to achieve state-of-the-art performance over models trained for specific quality settings.
arXiv Detail & Related papers (2020-04-17T00:10:08Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.