Recognition-Aware Learned Image Compression
- URL: http://arxiv.org/abs/2202.00198v1
- Date: Tue, 1 Feb 2022 03:33:51 GMT
- Title: Recognition-Aware Learned Image Compression
- Authors: Maxime Kawawa-Beaudan, Ryan Roggenkemper, Avideh Zakhor
- Abstract summary: We propose a recognition-aware learned compression method, which optimize a rate-distortion loss alongside a task-specific loss.
Our method achieves 26% higher recognition accuracy at equivalents compared to traditional methods such as BPG.
- Score: 0.5801044612920815
- License: http://creativecommons.org/licenses/by/4.0/
- Abstract: Learned image compression methods generally optimize a rate-distortion loss,
trading off improvements in visual distortion for added bitrate. Increasingly,
however, compressed imagery is used as an input to deep learning networks for
various tasks such as classification, object detection, and superresolution. We
propose a recognition-aware learned compression method, which optimizes a
rate-distortion loss alongside a task-specific loss, jointly learning
compression and recognition networks. We augment a hierarchical
autoencoder-based compression network with an EfficientNet recognition model
and use two hyperparameters to trade off between distortion, bitrate, and
recognition performance. We characterize the classification accuracy of our
proposed method as a function of bitrate and find that for low bitrates our
method achieves as much as 26% higher recognition accuracy at equivalent
bitrates compared to traditional methods such as Better Portable Graphics
(BPG).
Related papers
- Rate-Distortion-Cognition Controllable Versatile Neural Image Compression [47.72668401825835]
We propose a rate-distortion-cognition controllable versatile image compression method.
Our method yields satisfactory ICM performance and flexible Rate-DistortionCognition controlling.
arXiv Detail & Related papers (2024-07-16T13:17:51Z) - Enhancing the Rate-Distortion-Perception Flexibility of Learned Image
Codecs with Conditional Diffusion Decoders [7.485128109817576]
We show that conditional diffusion models can lead to promising results in the generative compression task when used as a decoder.
In this paper, we show that conditional diffusion models can lead to promising results in the generative compression task when used as a decoder.
arXiv Detail & Related papers (2024-03-05T11:48:35Z) - Semantic Ensemble Loss and Latent Refinement for High-Fidelity Neural Image Compression [58.618625678054826]
This study presents an enhanced neural compression method designed for optimal visual fidelity.
We have trained our model with a sophisticated semantic ensemble loss, integrating Charbonnier loss, perceptual loss, style loss, and a non-binary adversarial loss.
Our empirical findings demonstrate that this approach significantly improves the statistical fidelity of neural image compression.
arXiv Detail & Related papers (2024-01-25T08:11:27Z) - Progressive Learning with Visual Prompt Tuning for Variable-Rate Image
Compression [60.689646881479064]
We propose a progressive learning paradigm for transformer-based variable-rate image compression.
Inspired by visual prompt tuning, we use LPM to extract prompts for input images and hidden features at the encoder side and decoder side, respectively.
Our model outperforms all current variable image methods in terms of rate-distortion performance and approaches the state-of-the-art fixed image compression methods trained from scratch.
arXiv Detail & Related papers (2023-11-23T08:29:32Z) - You Can Mask More For Extremely Low-Bitrate Image Compression [80.7692466922499]
Learned image compression (LIC) methods have experienced significant progress during recent years.
LIC methods fail to explicitly explore the image structure and texture components crucial for image compression.
We present DA-Mask that samples visible patches based on the structure and texture of original images.
We propose a simple yet effective masked compression model (MCM), the first framework that unifies LIC and LIC end-to-end for extremely low-bitrate compression.
arXiv Detail & Related papers (2023-06-27T15:36:22Z) - Estimating the Resize Parameter in End-to-end Learned Image Compression [50.20567320015102]
We describe a search-free resizing framework that can further improve the rate-distortion tradeoff of recent learned image compression models.
Our results show that our new resizing parameter estimation framework can provide Bjontegaard-Delta rate (BD-rate) improvement of about 10% against leading perceptual quality engines.
arXiv Detail & Related papers (2022-04-26T01:35:02Z) - Neural JPEG: End-to-End Image Compression Leveraging a Standard JPEG
Encoder-Decoder [73.48927855855219]
We propose a system that learns to improve the encoding performance by enhancing its internal neural representations on both the encoder and decoder ends.
Experiments demonstrate that our approach successfully improves the rate-distortion performance over JPEG across various quality metrics.
arXiv Detail & Related papers (2022-01-27T20:20:03Z) - High-Fidelity Generative Image Compression [39.04379573099481]
We study how to combine Gene Adrial Networks and learned compression to obtain a state-of-the-art generative lossy compression system.
In particular, we investigate normalization layers, generator and discriminator architectures, training strategies, as well as perceptual losses.
arXiv Detail & Related papers (2020-06-17T16:21:10Z)
This list is automatically generated from the titles and abstracts of the papers in this site.
This site does not guarantee the quality of this site (including all information) and is not responsible for any consequences.